[ 492.456594] env[63088]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'linux_bridge' {{(pid=63088) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 492.456902] env[63088]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'noop' {{(pid=63088) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 492.457045] env[63088]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'ovs' {{(pid=63088) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 492.457331] env[63088]: INFO os_vif [-] Loaded VIF plugins: linux_bridge, noop, ovs [ 492.551738] env[63088]: DEBUG oslo_concurrency.processutils [-] Running cmd (subprocess): grep -F node.session.scan /sbin/iscsiadm {{(pid=63088) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:390}} [ 492.561590] env[63088]: DEBUG oslo_concurrency.processutils [-] CMD "grep -F node.session.scan /sbin/iscsiadm" returned: 0 in 0.010s {{(pid=63088) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:428}} [ 493.163785] env[63088]: INFO nova.virt.driver [None req-e7e47b96-710a-41e3-970a-b5a898e7ca8b None None] Loading compute driver 'vmwareapi.VMwareVCDriver' [ 493.235446] env[63088]: DEBUG oslo_concurrency.lockutils [-] Acquiring lock "oslo_vmware_api_lock" by "oslo_vmware.api.VMwareAPISession._create_session" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 493.235613] env[63088]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" acquired by "oslo_vmware.api.VMwareAPISession._create_session" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 493.235738] env[63088]: DEBUG oslo_vmware.service [-] Creating suds client with soap_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk' and wsdl_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk/vimService.wsdl' {{(pid=63088) __init__ /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:242}} [ 496.335752] env[63088]: DEBUG oslo_vmware.service [-] Invoking ServiceInstance.RetrieveServiceContent with opID=oslo.vmware-07f5c30c-ce28-4ed8-88df-943d630d511a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 496.351656] env[63088]: DEBUG oslo_vmware.api [-] Logging into host: vc1.osci.c.eu-de-1.cloud.sap. {{(pid=63088) _create_session /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:242}} [ 496.351812] env[63088]: DEBUG oslo_vmware.service [-] Invoking SessionManager.Login with opID=oslo.vmware-10ca299b-dc13-4331-b42a-1a5832183fd3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 496.378247] env[63088]: INFO oslo_vmware.api [-] Successfully established new session; session ID is 03483. [ 496.378374] env[63088]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" "released" by "oslo_vmware.api.VMwareAPISession._create_session" :: held 3.143s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 496.378892] env[63088]: INFO nova.virt.vmwareapi.driver [None req-e7e47b96-710a-41e3-970a-b5a898e7ca8b None None] VMware vCenter version: 7.0.3 [ 496.382215] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cd4dd33-d200-47df-934f-e56ce25c514e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 496.402920] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-616c4689-d3fa-49a0-8145-39da849f0f0d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 496.408648] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ed8a06b-2129-4846-9411-ad95531df67d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 496.415006] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2efff8c-dae8-4432-8069-cdf129eed251 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 496.427688] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43d29dea-f92a-4cdb-a1d2-7092e8db65da {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 496.433400] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d988c2cb-436b-4803-a8e1-b03755341181 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 496.462427] env[63088]: DEBUG oslo_vmware.service [-] Invoking ExtensionManager.FindExtension with opID=oslo.vmware-bb617dab-1da5-4917-9e78-3eeb85c9801a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 496.467279] env[63088]: DEBUG nova.virt.vmwareapi.driver [None req-e7e47b96-710a-41e3-970a-b5a898e7ca8b None None] Extension org.openstack.compute already exists. {{(pid=63088) _register_openstack_extension /opt/stack/nova/nova/virt/vmwareapi/driver.py:225}} [ 496.469926] env[63088]: INFO nova.compute.provider_config [None req-e7e47b96-710a-41e3-970a-b5a898e7ca8b None None] No provider configs found in /etc/nova/provider_config/. If files are present, ensure the Nova process has access. [ 496.972996] env[63088]: DEBUG nova.context [None req-e7e47b96-710a-41e3-970a-b5a898e7ca8b None None] Found 2 cells: 00000000-0000-0000-0000-000000000000(cell0),1b840a4f-6e0b-4bc7-b755-0ce3e5a83789(cell1) {{(pid=63088) load_cells /opt/stack/nova/nova/context.py:464}} [ 496.975127] env[63088]: DEBUG oslo_concurrency.lockutils [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 496.975348] env[63088]: DEBUG oslo_concurrency.lockutils [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 496.976016] env[63088]: DEBUG oslo_concurrency.lockutils [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 496.976450] env[63088]: DEBUG oslo_concurrency.lockutils [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] Acquiring lock "1b840a4f-6e0b-4bc7-b755-0ce3e5a83789" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 496.976703] env[63088]: DEBUG oslo_concurrency.lockutils [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] Lock "1b840a4f-6e0b-4bc7-b755-0ce3e5a83789" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 496.977731] env[63088]: DEBUG oslo_concurrency.lockutils [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] Lock "1b840a4f-6e0b-4bc7-b755-0ce3e5a83789" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 496.997877] env[63088]: INFO dbcounter [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] Registered counter for database nova_cell0 [ 497.006138] env[63088]: INFO dbcounter [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] Registered counter for database nova_cell1 [ 497.009558] env[63088]: DEBUG oslo_db.sqlalchemy.engines [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=63088) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 497.010183] env[63088]: DEBUG oslo_db.sqlalchemy.engines [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=63088) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 497.014602] env[63088]: ERROR nova.db.main.api [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 497.014602] env[63088]: result = function(*args, **kwargs) [ 497.014602] env[63088]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 497.014602] env[63088]: return func(*args, **kwargs) [ 497.014602] env[63088]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 497.014602] env[63088]: result = fn(*args, **kwargs) [ 497.014602] env[63088]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 497.014602] env[63088]: return f(*args, **kwargs) [ 497.014602] env[63088]: File "/opt/stack/nova/nova/objects/service.py", line 553, in _db_service_get_minimum_version [ 497.014602] env[63088]: return db.service_get_minimum_version(context, binaries) [ 497.014602] env[63088]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 497.014602] env[63088]: _check_db_access() [ 497.014602] env[63088]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 497.014602] env[63088]: stacktrace = ''.join(traceback.format_stack()) [ 497.014602] env[63088]: [ 497.015951] env[63088]: ERROR nova.db.main.api [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 497.015951] env[63088]: result = function(*args, **kwargs) [ 497.015951] env[63088]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 497.015951] env[63088]: return func(*args, **kwargs) [ 497.015951] env[63088]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 497.015951] env[63088]: result = fn(*args, **kwargs) [ 497.015951] env[63088]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 497.015951] env[63088]: return f(*args, **kwargs) [ 497.015951] env[63088]: File "/opt/stack/nova/nova/objects/service.py", line 553, in _db_service_get_minimum_version [ 497.015951] env[63088]: return db.service_get_minimum_version(context, binaries) [ 497.015951] env[63088]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 497.015951] env[63088]: _check_db_access() [ 497.015951] env[63088]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 497.015951] env[63088]: stacktrace = ''.join(traceback.format_stack()) [ 497.015951] env[63088]: [ 497.016566] env[63088]: WARNING nova.objects.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] Failed to get minimum service version for cell 00000000-0000-0000-0000-000000000000 [ 497.016566] env[63088]: WARNING nova.objects.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] Failed to get minimum service version for cell 1b840a4f-6e0b-4bc7-b755-0ce3e5a83789 [ 497.017078] env[63088]: DEBUG oslo_concurrency.lockutils [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] Acquiring lock "singleton_lock" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 497.017078] env[63088]: DEBUG oslo_concurrency.lockutils [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] Acquired lock "singleton_lock" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 497.017337] env[63088]: DEBUG oslo_concurrency.lockutils [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] Releasing lock "singleton_lock" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 497.017648] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] Full set of CONF: {{(pid=63088) _wait_for_exit_or_signal /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/service.py:363}} [ 497.017790] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] ******************************************************************************** {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2806}} [ 497.017917] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] Configuration options gathered from: {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2807}} [ 497.018062] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] command line args: ['--config-file', '/etc/nova/nova.conf', '--config-file', '/etc/nova/nova-cpu-common.conf', '--config-file', '/etc/nova/nova-cpu-1.conf'] {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2808}} [ 497.018256] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] config files: ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2809}} [ 497.018383] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] ================================================================================ {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2811}} [ 497.018594] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] allow_resize_to_same_host = True {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.018765] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] arq_binding_timeout = 300 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.018895] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] backdoor_port = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.019029] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] backdoor_socket = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.019198] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] block_device_allocate_retries = 60 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.019359] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] block_device_allocate_retries_interval = 3 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.019523] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cert = self.pem {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.019689] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] compute_driver = vmwareapi.VMwareVCDriver {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.019856] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] compute_monitors = [] {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.020154] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] config_dir = [] {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.020248] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] config_drive_format = iso9660 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.020334] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] config_file = ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.020481] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] config_source = [] {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.020641] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] console_host = devstack {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.020805] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] control_exchange = nova {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.020962] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cpu_allocation_ratio = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.021133] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] daemon = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.021298] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] debug = True {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.021450] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] default_access_ip_network_name = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.021644] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] default_availability_zone = nova {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.021871] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] default_ephemeral_format = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.022067] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] default_green_pool_size = 1000 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.022313] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'glanceclient=WARN', 'oslo.privsep.daemon=INFO'] {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.022480] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] default_schedule_zone = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.022637] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] disk_allocation_ratio = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.022797] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] enable_new_services = True {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.022976] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] enabled_apis = ['osapi_compute'] {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.023156] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] enabled_ssl_apis = [] {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.023318] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] flat_injected = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.023474] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] force_config_drive = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.023631] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] force_raw_images = True {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.023799] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] graceful_shutdown_timeout = 5 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.023959] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] heal_instance_info_cache_interval = 60 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.024181] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] host = cpu-1 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.024357] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] initial_cpu_allocation_ratio = 4.0 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.024545] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] initial_disk_allocation_ratio = 1.0 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.024784] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] initial_ram_allocation_ratio = 1.0 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.025048] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] injected_network_template = /opt/stack/nova/nova/virt/interfaces.template {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.025224] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] instance_build_timeout = 0 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.025390] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] instance_delete_interval = 300 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.025557] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] instance_format = [instance: %(uuid)s] {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.025728] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] instance_name_template = instance-%08x {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.025889] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] instance_usage_audit = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.026070] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] instance_usage_audit_period = month {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.026241] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] instance_uuid_format = [instance: %(uuid)s] {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.026407] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] instances_path = /opt/stack/data/nova/instances {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.026607] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] internal_service_availability_zone = internal {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.026772] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] key = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.026954] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] live_migration_retry_count = 30 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.027164] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] log_color = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.027335] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] log_config_append = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.027530] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] log_date_format = %Y-%m-%d %H:%M:%S {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.027729] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] log_dir = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.027941] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] log_file = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.028091] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] log_options = True {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.028262] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] log_rotate_interval = 1 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.028431] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] log_rotate_interval_type = days {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.028598] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] log_rotation_type = none {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.028729] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] logging_context_format_string = %(color)s%(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(project_name)s %(user_name)s%(color)s] %(instance)s%(color)s%(message)s {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.028855] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] logging_debug_format_suffix = {{(pid=%(process)d) %(funcName)s %(pathname)s:%(lineno)d}} {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.029043] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] logging_default_format_string = %(color)s%(levelname)s %(name)s [-%(color)s] %(instance)s%(color)s%(message)s {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.029218] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] logging_exception_prefix = ERROR %(name)s %(instance)s {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.029345] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.029505] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] long_rpc_timeout = 1800 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.029666] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] max_concurrent_builds = 10 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.029823] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] max_concurrent_live_migrations = 1 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.029979] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] max_concurrent_snapshots = 5 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.030153] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] max_local_block_devices = 3 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.030311] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] max_logfile_count = 30 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.030468] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] max_logfile_size_mb = 200 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.030633] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] maximum_instance_delete_attempts = 5 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.030797] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] metadata_listen = 0.0.0.0 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.030965] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] metadata_listen_port = 8775 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.031148] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] metadata_workers = 2 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.031311] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] migrate_max_retries = -1 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.031475] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] mkisofs_cmd = genisoimage {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.031680] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] my_block_storage_ip = 10.180.1.21 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.031815] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] my_ip = 10.180.1.21 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.031977] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] network_allocate_retries = 0 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.032168] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] non_inheritable_image_properties = ['cache_in_nova', 'bittorrent'] {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.032337] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] osapi_compute_listen = 0.0.0.0 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.032499] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] osapi_compute_listen_port = 8774 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.032668] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] osapi_compute_unique_server_name_scope = {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.032837] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] osapi_compute_workers = 2 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.033011] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] password_length = 12 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.033175] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] periodic_enable = True {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.033338] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] periodic_fuzzy_delay = 60 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.033506] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] pointer_model = usbtablet {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.033679] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] preallocate_images = none {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.033832] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] publish_errors = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.033961] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] pybasedir = /opt/stack/nova {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.034168] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] ram_allocation_ratio = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.034289] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] rate_limit_burst = 0 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.034455] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] rate_limit_except_level = CRITICAL {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.034637] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] rate_limit_interval = 0 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.034801] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] reboot_timeout = 0 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.034960] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] reclaim_instance_interval = 0 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.035129] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] record = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.035300] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] reimage_timeout_per_gb = 60 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.035466] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] report_interval = 120 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.035626] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] rescue_timeout = 0 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.035786] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] reserved_host_cpus = 0 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.035944] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] reserved_host_disk_mb = 0 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.036114] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] reserved_host_memory_mb = 512 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.036277] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] reserved_huge_pages = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.036438] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] resize_confirm_window = 0 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.036625] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] resize_fs_using_block_device = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.036793] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] resume_guests_state_on_host_boot = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.036963] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] rootwrap_config = /etc/nova/rootwrap.conf {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.037138] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] rpc_response_timeout = 60 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.037299] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] run_external_periodic_tasks = True {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.037472] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] running_deleted_instance_action = reap {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.037630] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] running_deleted_instance_poll_interval = 1800 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.037789] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] running_deleted_instance_timeout = 0 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.037946] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] scheduler_instance_sync_interval = 120 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.038127] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] service_down_time = 720 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.038297] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] servicegroup_driver = db {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.038452] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] shell_completion = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.038610] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] shelved_offload_time = 0 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.038767] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] shelved_poll_interval = 3600 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.038961] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] shutdown_timeout = 0 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.039145] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] source_is_ipv6 = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.039305] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] ssl_only = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.039548] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] state_path = /opt/stack/data/n-cpu-1 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.039718] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] sync_power_state_interval = 600 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.039879] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] sync_power_state_pool_size = 1000 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.040059] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] syslog_log_facility = LOG_USER {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.040221] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] tempdir = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.040382] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] timeout_nbd = 10 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.040550] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] transport_url = **** {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.040713] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] update_resources_interval = 0 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.040870] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] use_cow_images = True {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.041033] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] use_eventlog = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.041195] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] use_journal = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.041350] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] use_json = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.041505] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] use_rootwrap_daemon = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.041662] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] use_stderr = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.041817] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] use_syslog = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.041966] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] vcpu_pin_set = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.042143] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] vif_plugging_is_fatal = True {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.042308] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] vif_plugging_timeout = 300 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.042470] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] virt_mkfs = [] {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.042629] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] volume_usage_poll_interval = 0 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.042787] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] watch_log_file = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.042948] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] web = /usr/share/spice-html5 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 497.043141] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_concurrency.disable_process_locking = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.043712] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_concurrency.lock_path = /opt/stack/data/n-cpu-1 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.043906] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_messaging_metrics.metrics_buffer_size = 1000 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.044093] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_messaging_metrics.metrics_enabled = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.044275] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_messaging_metrics.metrics_process_name = {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.044450] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_messaging_metrics.metrics_socket_file = /var/tmp/metrics_collector.sock {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.044618] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_messaging_metrics.metrics_thread_stop_timeout = 10 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.044806] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] api.auth_strategy = keystone {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.044975] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] api.compute_link_prefix = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.045204] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] api.config_drive_skip_versions = 1.0 2007-01-19 2007-03-01 2007-08-29 2007-10-10 2007-12-15 2008-02-01 2008-09-01 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.045660] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] api.dhcp_domain = novalocal {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.045874] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] api.enable_instance_password = True {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.046067] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] api.glance_link_prefix = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.046246] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] api.instance_list_cells_batch_fixed_size = 100 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.046415] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] api.instance_list_cells_batch_strategy = distributed {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.046627] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] api.instance_list_per_project_cells = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.046817] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] api.list_records_by_skipping_down_cells = True {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.046986] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] api.local_metadata_per_cell = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.047174] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] api.max_limit = 1000 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.047347] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] api.metadata_cache_expiration = 15 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.047553] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] api.neutron_default_tenant_id = default {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.047754] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] api.response_validation = warn {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.047935] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] api.use_neutron_default_nets = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.048164] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] api.vendordata_dynamic_connect_timeout = 5 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.048455] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] api.vendordata_dynamic_failure_fatal = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.048602] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] api.vendordata_dynamic_read_timeout = 5 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.048802] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] api.vendordata_dynamic_ssl_certfile = {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.048986] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] api.vendordata_dynamic_targets = [] {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.049195] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] api.vendordata_jsonfile_path = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.049435] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] api.vendordata_providers = ['StaticJSON'] {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.049640] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cache.backend = dogpile.cache.memcached {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.049813] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cache.backend_argument = **** {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.049986] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cache.config_prefix = cache.oslo {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.050175] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cache.dead_timeout = 60.0 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.050341] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cache.debug_cache_backend = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.050507] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cache.enable_retry_client = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.050670] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cache.enable_socket_keepalive = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.050841] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cache.enabled = True {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.051010] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cache.enforce_fips_mode = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.051183] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cache.expiration_time = 600 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.051347] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cache.hashclient_retry_attempts = 2 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.051549] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cache.hashclient_retry_delay = 1.0 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.051764] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cache.memcache_dead_retry = 300 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.051933] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cache.memcache_password = **** {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.052115] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cache.memcache_pool_connection_get_timeout = 10 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.052284] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cache.memcache_pool_flush_on_reconnect = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.052450] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cache.memcache_pool_maxsize = 10 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.052613] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cache.memcache_pool_unused_timeout = 60 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.052779] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cache.memcache_sasl_enabled = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.052960] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cache.memcache_servers = ['localhost:11211'] {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.053150] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cache.memcache_socket_timeout = 1.0 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.053314] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cache.memcache_username = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.053483] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cache.proxies = [] {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.053652] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cache.redis_db = 0 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.053810] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cache.redis_password = **** {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.053984] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cache.redis_sentinel_service_name = mymaster {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.054213] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cache.redis_sentinels = ['localhost:26379'] {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.054396] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cache.redis_server = localhost:6379 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.054566] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cache.redis_socket_timeout = 1.0 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.054728] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cache.redis_username = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.054890] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cache.retry_attempts = 2 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.055069] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cache.retry_delay = 0.0 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.055238] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cache.socket_keepalive_count = 1 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.055400] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cache.socket_keepalive_idle = 1 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.055562] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cache.socket_keepalive_interval = 1 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.055721] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cache.tls_allowed_ciphers = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.055876] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cache.tls_cafile = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.056057] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cache.tls_certfile = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.056233] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cache.tls_enabled = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.056391] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cache.tls_keyfile = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.056599] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cinder.auth_section = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.056780] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cinder.auth_type = password {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.056944] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cinder.cafile = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.057134] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cinder.catalog_info = volumev3::publicURL {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.057297] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cinder.certfile = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.057470] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cinder.collect_timing = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.057664] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cinder.cross_az_attach = True {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.057837] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cinder.debug = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.057997] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cinder.endpoint_template = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.058179] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cinder.http_retries = 3 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.058343] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cinder.insecure = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.058500] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cinder.keyfile = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.058673] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cinder.os_region_name = RegionOne {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.058834] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cinder.split_loggers = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.058991] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cinder.timeout = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.059180] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] compute.consecutive_build_service_disable_threshold = 10 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.059345] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] compute.cpu_dedicated_set = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.059506] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] compute.cpu_shared_set = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.059673] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] compute.image_type_exclude_list = [] {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.059848] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] compute.live_migration_wait_for_vif_plug = True {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.059994] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] compute.max_concurrent_disk_ops = 0 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.060170] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] compute.max_disk_devices_to_attach = -1 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.060333] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] compute.packing_host_numa_cells_allocation_strategy = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.060501] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] compute.provider_config_location = /etc/nova/provider_config/ {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.060663] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] compute.resource_provider_association_refresh = 300 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.060822] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] compute.sharing_providers_max_uuids_per_request = 200 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.060982] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] compute.shutdown_retry_interval = 10 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.061176] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] compute.vmdk_allowed_types = ['streamOptimized', 'monolithicSparse'] {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.061359] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] conductor.workers = 2 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.061541] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] console.allowed_origins = [] {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.061703] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] console.ssl_ciphers = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.061871] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] console.ssl_minimum_version = default {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.062050] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] consoleauth.enforce_session_timeout = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.062225] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] consoleauth.token_ttl = 600 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.062400] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cyborg.cafile = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.062558] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cyborg.certfile = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.062721] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cyborg.collect_timing = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.062880] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cyborg.connect_retries = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.063054] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cyborg.connect_retry_delay = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.063221] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cyborg.endpoint_override = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.063382] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cyborg.insecure = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.063538] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cyborg.keyfile = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.063701] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cyborg.max_version = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.063924] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cyborg.min_version = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.064114] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cyborg.region_name = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.064282] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cyborg.retriable_status_codes = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.064441] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cyborg.service_name = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.064611] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cyborg.service_type = accelerator {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.064776] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cyborg.split_loggers = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.064937] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cyborg.status_code_retries = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.065109] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cyborg.status_code_retry_delay = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.065269] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cyborg.timeout = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.065451] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cyborg.valid_interfaces = ['internal', 'public'] {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.065611] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] cyborg.version = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.065791] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] database.backend = sqlalchemy {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.065962] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] database.connection = **** {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.066145] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] database.connection_debug = 0 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.066317] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] database.connection_parameters = {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.066497] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] database.connection_recycle_time = 3600 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.066681] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] database.connection_trace = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.066848] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] database.db_inc_retry_interval = True {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.067023] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] database.db_max_retries = 20 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.067193] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] database.db_max_retry_interval = 10 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.067356] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] database.db_retry_interval = 1 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.067546] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] database.max_overflow = 50 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.067726] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] database.max_pool_size = 5 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.067892] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] database.max_retries = 10 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.068074] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] database.mysql_sql_mode = TRADITIONAL {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.068239] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] database.mysql_wsrep_sync_wait = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.068398] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] database.pool_timeout = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.068561] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] database.retry_interval = 10 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.068722] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] database.slave_connection = **** {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.068882] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] database.sqlite_synchronous = True {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.069054] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] database.use_db_reconnect = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.069239] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] api_database.backend = sqlalchemy {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.069410] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] api_database.connection = **** {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.069575] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] api_database.connection_debug = 0 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.069749] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] api_database.connection_parameters = {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.069911] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] api_database.connection_recycle_time = 3600 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.070085] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] api_database.connection_trace = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.070252] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] api_database.db_inc_retry_interval = True {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.070416] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] api_database.db_max_retries = 20 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.070581] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] api_database.db_max_retry_interval = 10 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.070746] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] api_database.db_retry_interval = 1 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.070905] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] api_database.max_overflow = 50 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.071078] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] api_database.max_pool_size = 5 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.071246] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] api_database.max_retries = 10 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.071415] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] api_database.mysql_sql_mode = TRADITIONAL {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.071574] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] api_database.mysql_wsrep_sync_wait = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.071735] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] api_database.pool_timeout = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.071897] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] api_database.retry_interval = 10 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.072067] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] api_database.slave_connection = **** {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.072233] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] api_database.sqlite_synchronous = True {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.072409] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] devices.enabled_mdev_types = [] {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.072586] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] ephemeral_storage_encryption.cipher = aes-xts-plain64 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.072764] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] ephemeral_storage_encryption.default_format = luks {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.072924] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] ephemeral_storage_encryption.enabled = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.073102] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] ephemeral_storage_encryption.key_size = 512 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.073277] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] glance.api_servers = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.073441] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] glance.cafile = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.073600] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] glance.certfile = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.073763] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] glance.collect_timing = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.073920] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] glance.connect_retries = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.074091] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] glance.connect_retry_delay = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.074255] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] glance.debug = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.074421] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] glance.default_trusted_certificate_ids = [] {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.074584] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] glance.enable_certificate_validation = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.074745] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] glance.enable_rbd_download = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.074904] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] glance.endpoint_override = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.075080] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] glance.insecure = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.075248] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] glance.keyfile = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.075408] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] glance.max_version = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.075564] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] glance.min_version = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.075732] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] glance.num_retries = 3 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.075899] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] glance.rbd_ceph_conf = {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.076074] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] glance.rbd_connect_timeout = 5 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.076247] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] glance.rbd_pool = {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.076415] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] glance.rbd_user = {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.076599] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] glance.region_name = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.076769] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] glance.retriable_status_codes = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.076929] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] glance.service_name = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.077110] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] glance.service_type = image {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.077273] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] glance.split_loggers = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.077430] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] glance.status_code_retries = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.077616] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] glance.status_code_retry_delay = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.077788] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] glance.timeout = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.077975] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] glance.valid_interfaces = ['internal', 'public'] {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.078156] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] glance.verify_glance_signatures = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.078320] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] glance.version = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.078487] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] guestfs.debug = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.078655] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] mks.enabled = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.079043] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] mks.mksproxy_base_url = http://127.0.0.1:6090/ {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.079249] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] image_cache.manager_interval = 2400 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.079422] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] image_cache.precache_concurrency = 1 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.079596] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] image_cache.remove_unused_base_images = True {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.079770] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] image_cache.remove_unused_original_minimum_age_seconds = 86400 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.079941] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] image_cache.remove_unused_resized_minimum_age_seconds = 3600 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.080131] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] image_cache.subdirectory_name = _base {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.080313] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] ironic.api_max_retries = 60 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.080480] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] ironic.api_retry_interval = 2 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.080643] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] ironic.auth_section = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.080810] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] ironic.auth_type = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.080971] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] ironic.cafile = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.081166] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] ironic.certfile = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.081310] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] ironic.collect_timing = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.081475] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] ironic.conductor_group = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.081636] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] ironic.connect_retries = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.081797] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] ironic.connect_retry_delay = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.081954] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] ironic.endpoint_override = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.082129] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] ironic.insecure = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.082290] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] ironic.keyfile = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.082449] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] ironic.max_version = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.082608] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] ironic.min_version = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.082774] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] ironic.peer_list = [] {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.082932] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] ironic.region_name = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.083104] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] ironic.retriable_status_codes = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.083273] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] ironic.serial_console_state_timeout = 10 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.083433] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] ironic.service_name = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.083602] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] ironic.service_type = baremetal {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.083763] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] ironic.shard = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.083923] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] ironic.split_loggers = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.084093] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] ironic.status_code_retries = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.084256] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] ironic.status_code_retry_delay = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.084416] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] ironic.timeout = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.084631] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] ironic.valid_interfaces = ['internal', 'public'] {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.084758] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] ironic.version = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.084941] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] key_manager.backend = nova.keymgr.conf_key_mgr.ConfKeyManager {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.085126] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] key_manager.fixed_key = **** {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.085311] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] barbican.auth_endpoint = http://localhost/identity/v3 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.085475] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] barbican.barbican_api_version = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.085632] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] barbican.barbican_endpoint = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.085803] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] barbican.barbican_endpoint_type = public {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.085960] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] barbican.barbican_region_name = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.086133] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] barbican.cafile = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.086295] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] barbican.certfile = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.086456] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] barbican.collect_timing = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.086644] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] barbican.insecure = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.086807] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] barbican.keyfile = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.086970] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] barbican.number_of_retries = 60 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.087145] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] barbican.retry_delay = 1 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.087308] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] barbican.send_service_user_token = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.087476] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] barbican.split_loggers = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.087654] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] barbican.timeout = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.087844] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] barbican.verify_ssl = True {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.087973] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] barbican.verify_ssl_path = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.088157] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] barbican_service_user.auth_section = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.088323] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] barbican_service_user.auth_type = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.088484] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] barbican_service_user.cafile = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.088643] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] barbican_service_user.certfile = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.088807] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] barbican_service_user.collect_timing = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.088996] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] barbican_service_user.insecure = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.089178] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] barbican_service_user.keyfile = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.089343] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] barbican_service_user.split_loggers = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.089502] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] barbican_service_user.timeout = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.089672] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] vault.approle_role_id = **** {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.089830] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] vault.approle_secret_id = **** {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.090005] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] vault.kv_mountpoint = secret {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.090175] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] vault.kv_path = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.090340] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] vault.kv_version = 2 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.090501] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] vault.namespace = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.090659] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] vault.root_token_id = **** {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.090815] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] vault.ssl_ca_crt_file = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.090978] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] vault.timeout = 60.0 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.091151] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] vault.use_ssl = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.091324] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] vault.vault_url = http://127.0.0.1:8200 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.091492] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] keystone.auth_section = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.091659] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] keystone.auth_type = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.091816] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] keystone.cafile = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.091973] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] keystone.certfile = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.092151] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] keystone.collect_timing = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.092311] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] keystone.connect_retries = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.092470] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] keystone.connect_retry_delay = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.092628] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] keystone.endpoint_override = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.092791] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] keystone.insecure = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.092946] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] keystone.keyfile = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.093127] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] keystone.max_version = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.093289] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] keystone.min_version = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.093447] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] keystone.region_name = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.093603] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] keystone.retriable_status_codes = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.093759] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] keystone.service_name = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.093927] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] keystone.service_type = identity {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.094101] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] keystone.split_loggers = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.094264] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] keystone.status_code_retries = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.094421] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] keystone.status_code_retry_delay = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.094579] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] keystone.timeout = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.094760] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] keystone.valid_interfaces = ['internal', 'public'] {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.094919] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] keystone.version = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.095132] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.connection_uri = {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.095297] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.cpu_mode = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.095465] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.cpu_model_extra_flags = [] {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.095636] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.cpu_models = [] {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.095806] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.cpu_power_governor_high = performance {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.095972] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.cpu_power_governor_low = powersave {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.096145] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.cpu_power_management = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.096314] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.cpu_power_management_strategy = cpu_state {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.096480] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.device_detach_attempts = 8 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.096670] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.device_detach_timeout = 20 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.096837] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.disk_cachemodes = [] {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.096996] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.disk_prefix = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.097175] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.enabled_perf_events = [] {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.097339] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.file_backed_memory = 0 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.097521] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.gid_maps = [] {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.097694] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.hw_disk_discard = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.097854] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.hw_machine_type = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.098033] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.images_rbd_ceph_conf = {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.098205] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.images_rbd_glance_copy_poll_interval = 15 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.098368] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.images_rbd_glance_copy_timeout = 600 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.098542] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.images_rbd_glance_store_name = {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.098713] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.images_rbd_pool = rbd {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.098910] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.images_type = default {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.099100] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.images_volume_group = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.099268] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.inject_key = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.099433] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.inject_partition = -2 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.099596] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.inject_password = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.099763] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.iscsi_iface = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.099925] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.iser_use_multipath = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.100100] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.live_migration_bandwidth = 0 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.100270] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.live_migration_completion_timeout = 800 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.100433] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.live_migration_downtime = 500 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.100596] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.live_migration_downtime_delay = 75 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.100760] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.live_migration_downtime_steps = 10 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.100922] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.live_migration_inbound_addr = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.101093] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.live_migration_permit_auto_converge = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.101257] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.live_migration_permit_post_copy = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.101417] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.live_migration_scheme = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.101586] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.live_migration_timeout_action = abort {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.101753] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.live_migration_uri = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.101920] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.live_migration_with_native_tls = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.102096] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.max_queues = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.102267] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.mem_stats_period_seconds = 10 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.102501] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.migration_inbound_addr = 10.180.1.21 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.102670] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.nfs_mount_options = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.102970] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.nfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.103164] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.num_aoe_discover_tries = 3 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.103332] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.num_iser_scan_tries = 5 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.103496] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.num_memory_encrypted_guests = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.103662] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.num_nvme_discover_tries = 5 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.103829] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.num_pcie_ports = 0 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.103996] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.num_volume_scan_tries = 5 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.104178] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.pmem_namespaces = [] {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.104340] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.quobyte_client_cfg = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.104636] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.quobyte_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.104826] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.rbd_connect_timeout = 5 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.104975] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.rbd_destroy_volume_retries = 12 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.105153] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.rbd_destroy_volume_retry_interval = 5 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.105315] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.rbd_secret_uuid = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.105472] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.rbd_user = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.105642] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.realtime_scheduler_priority = 1 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.105814] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.remote_filesystem_transport = ssh {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.105978] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.rescue_image_id = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.106151] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.rescue_kernel_id = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.106310] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.rescue_ramdisk_id = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.106481] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.rng_dev_path = /dev/urandom {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.106671] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.rx_queue_size = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.106847] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.smbfs_mount_options = {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.107180] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.smbfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.107363] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.snapshot_compression = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.107549] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.snapshot_image_format = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.107785] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.snapshots_directory = /opt/stack/data/nova/instances/snapshots {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.107955] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.sparse_logical_volumes = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.108138] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.swtpm_enabled = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.108312] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.swtpm_group = tss {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.108484] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.swtpm_user = tss {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.108658] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.sysinfo_serial = unique {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.108822] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.tb_cache_size = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.108983] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.tx_queue_size = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.109165] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.uid_maps = [] {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.109331] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.use_virtio_for_bridges = True {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.109500] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.virt_type = kvm {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.109673] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.volume_clear = zero {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.109836] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.volume_clear_size = 0 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.110007] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.volume_use_multipath = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.110176] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.vzstorage_cache_path = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.110346] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.vzstorage_log_path = /var/log/vstorage/%(cluster_name)s/nova.log.gz {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.110523] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.vzstorage_mount_group = qemu {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.110686] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.vzstorage_mount_opts = [] {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.110858] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.vzstorage_mount_perms = 0770 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.111146] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.vzstorage_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.111327] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.vzstorage_mount_user = stack {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.111501] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] libvirt.wait_soft_reboot_seconds = 120 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.111678] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] neutron.auth_section = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.111854] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] neutron.auth_type = password {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.112028] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] neutron.cafile = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.112197] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] neutron.certfile = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.112362] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] neutron.collect_timing = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.112526] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] neutron.connect_retries = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.112688] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] neutron.connect_retry_delay = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.112858] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] neutron.default_floating_pool = public {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.113027] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] neutron.endpoint_override = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.113199] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] neutron.extension_sync_interval = 600 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.113365] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] neutron.http_retries = 3 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.113531] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] neutron.insecure = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.113692] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] neutron.keyfile = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.113849] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] neutron.max_version = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.114028] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] neutron.metadata_proxy_shared_secret = **** {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.114196] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] neutron.min_version = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.114366] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] neutron.ovs_bridge = br-int {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.114534] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] neutron.physnets = [] {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.114705] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] neutron.region_name = RegionOne {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.114918] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] neutron.retriable_status_codes = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.115043] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] neutron.service_metadata_proxy = True {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.115208] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] neutron.service_name = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.115379] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] neutron.service_type = network {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.115541] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] neutron.split_loggers = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.115701] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] neutron.status_code_retries = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.115859] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] neutron.status_code_retry_delay = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.116025] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] neutron.timeout = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.116212] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] neutron.valid_interfaces = ['internal', 'public'] {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.116375] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] neutron.version = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.116568] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] notifications.bdms_in_notifications = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.116756] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] notifications.default_level = INFO {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.116933] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] notifications.notification_format = unversioned {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.117111] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] notifications.notify_on_state_change = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.117292] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] notifications.versioned_notifications_topics = ['versioned_notifications'] {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.117474] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] pci.alias = [] {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.117711] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] pci.device_spec = [] {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.117894] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] pci.report_in_placement = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.118083] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] placement.auth_section = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.118261] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] placement.auth_type = password {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.118430] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] placement.auth_url = http://10.180.1.21/identity {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.118591] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] placement.cafile = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.118751] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] placement.certfile = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.118939] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] placement.collect_timing = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.119120] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] placement.connect_retries = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.119283] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] placement.connect_retry_delay = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.119439] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] placement.default_domain_id = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.119594] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] placement.default_domain_name = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.119815] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] placement.domain_id = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.120053] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] placement.domain_name = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.120235] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] placement.endpoint_override = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.120400] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] placement.insecure = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.120560] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] placement.keyfile = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.120720] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] placement.max_version = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.120877] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] placement.min_version = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.121053] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] placement.password = **** {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.121216] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] placement.project_domain_id = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.121382] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] placement.project_domain_name = Default {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.121550] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] placement.project_id = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.121721] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] placement.project_name = service {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.121889] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] placement.region_name = RegionOne {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.122058] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] placement.retriable_status_codes = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.122221] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] placement.service_name = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.122388] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] placement.service_type = placement {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.122552] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] placement.split_loggers = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.122710] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] placement.status_code_retries = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.122866] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] placement.status_code_retry_delay = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.123035] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] placement.system_scope = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.123198] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] placement.timeout = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.123356] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] placement.trust_id = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.123513] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] placement.user_domain_id = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.123682] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] placement.user_domain_name = Default {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.123839] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] placement.user_id = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.124026] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] placement.username = nova {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.124200] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] placement.valid_interfaces = ['internal', 'public'] {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.124360] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] placement.version = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.124537] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] quota.cores = 20 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.124700] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] quota.count_usage_from_placement = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.124871] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] quota.driver = nova.quota.DbQuotaDriver {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.125059] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] quota.injected_file_content_bytes = 10240 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.125229] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] quota.injected_file_path_length = 255 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.125393] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] quota.injected_files = 5 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.125561] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] quota.instances = 10 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.125725] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] quota.key_pairs = 100 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.125888] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] quota.metadata_items = 128 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.126063] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] quota.ram = 51200 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.126229] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] quota.recheck_quota = True {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.126396] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] quota.server_group_members = 10 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.126596] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] quota.server_groups = 10 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.126784] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] scheduler.discover_hosts_in_cells_interval = -1 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.126987] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] scheduler.enable_isolated_aggregate_filtering = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.127183] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] scheduler.image_metadata_prefilter = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.127350] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] scheduler.limit_tenants_to_placement_aggregate = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.127523] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] scheduler.max_attempts = 3 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.127684] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] scheduler.max_placement_results = 1000 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.127849] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] scheduler.placement_aggregate_required_for_tenants = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.128019] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] scheduler.query_placement_for_image_type_support = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.128184] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] scheduler.query_placement_for_routed_network_aggregates = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.128360] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] scheduler.workers = 2 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.128534] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] filter_scheduler.aggregate_image_properties_isolation_namespace = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.128707] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] filter_scheduler.aggregate_image_properties_isolation_separator = . {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.128910] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] filter_scheduler.available_filters = ['nova.scheduler.filters.all_filters'] {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.129106] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] filter_scheduler.build_failure_weight_multiplier = 1000000.0 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.129279] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] filter_scheduler.cpu_weight_multiplier = 1.0 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.129445] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] filter_scheduler.cross_cell_move_weight_multiplier = 1000000.0 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.129610] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] filter_scheduler.disk_weight_multiplier = 1.0 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.129802] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] filter_scheduler.enabled_filters = ['ComputeFilter', 'ComputeCapabilitiesFilter', 'ImagePropertiesFilter', 'ServerGroupAntiAffinityFilter', 'ServerGroupAffinityFilter', 'SameHostFilter', 'DifferentHostFilter'] {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.129972] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] filter_scheduler.host_subset_size = 1 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.130154] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] filter_scheduler.hypervisor_version_weight_multiplier = 1.0 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.130316] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] filter_scheduler.image_properties_default_architecture = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.130479] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] filter_scheduler.io_ops_weight_multiplier = -1.0 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.130646] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] filter_scheduler.isolated_hosts = [] {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.130810] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] filter_scheduler.isolated_images = [] {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.130978] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] filter_scheduler.max_instances_per_host = 50 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.131157] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] filter_scheduler.max_io_ops_per_host = 8 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.131322] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] filter_scheduler.num_instances_weight_multiplier = 0.0 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.131487] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] filter_scheduler.pci_in_placement = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.131652] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] filter_scheduler.pci_weight_multiplier = 1.0 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.131813] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] filter_scheduler.ram_weight_multiplier = 1.0 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.131976] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] filter_scheduler.restrict_isolated_hosts_to_isolated_images = True {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.132149] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] filter_scheduler.shuffle_best_same_weighed_hosts = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.132314] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] filter_scheduler.soft_affinity_weight_multiplier = 1.0 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.132474] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] filter_scheduler.soft_anti_affinity_weight_multiplier = 1.0 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.132637] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] filter_scheduler.track_instance_changes = True {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.132810] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] filter_scheduler.weight_classes = ['nova.scheduler.weights.all_weighers'] {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.132976] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] metrics.required = True {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.133154] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] metrics.weight_multiplier = 1.0 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.133317] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] metrics.weight_of_unavailable = -10000.0 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.133482] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] metrics.weight_setting = [] {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.133804] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] serial_console.base_url = ws://127.0.0.1:6083/ {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.133974] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] serial_console.enabled = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.134166] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] serial_console.port_range = 10000:20000 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.134339] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] serial_console.proxyclient_address = 127.0.0.1 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.134509] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] serial_console.serialproxy_host = 0.0.0.0 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.134677] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] serial_console.serialproxy_port = 6083 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.134843] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] service_user.auth_section = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.135061] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] service_user.auth_type = password {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.135190] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] service_user.cafile = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.135348] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] service_user.certfile = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.135512] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] service_user.collect_timing = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.135676] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] service_user.insecure = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.135833] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] service_user.keyfile = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.136007] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] service_user.send_service_user_token = True {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.136185] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] service_user.split_loggers = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.136346] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] service_user.timeout = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.136533] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] spice.agent_enabled = True {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.136709] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] spice.enabled = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.137030] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] spice.html5proxy_base_url = http://127.0.0.1:6082/spice_auto.html {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.137239] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] spice.html5proxy_host = 0.0.0.0 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.137412] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] spice.html5proxy_port = 6082 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.137572] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] spice.image_compression = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.137735] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] spice.jpeg_compression = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.137892] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] spice.playback_compression = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.138066] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] spice.require_secure = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.138238] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] spice.server_listen = 127.0.0.1 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.138406] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] spice.server_proxyclient_address = 127.0.0.1 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.138565] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] spice.streaming_mode = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.138722] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] spice.zlib_compression = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.138909] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] upgrade_levels.baseapi = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.139101] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] upgrade_levels.compute = auto {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.139264] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] upgrade_levels.conductor = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.139422] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] upgrade_levels.scheduler = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.139587] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] vendordata_dynamic_auth.auth_section = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.139749] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] vendordata_dynamic_auth.auth_type = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.139904] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] vendordata_dynamic_auth.cafile = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.140071] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] vendordata_dynamic_auth.certfile = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.140235] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] vendordata_dynamic_auth.collect_timing = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.140393] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] vendordata_dynamic_auth.insecure = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.140549] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] vendordata_dynamic_auth.keyfile = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.140712] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] vendordata_dynamic_auth.split_loggers = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.140869] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] vendordata_dynamic_auth.timeout = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.141051] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] vmware.api_retry_count = 10 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.141214] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] vmware.ca_file = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.141382] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] vmware.cache_prefix = devstack-image-cache {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.141548] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] vmware.cluster_name = testcl1 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.141713] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] vmware.connection_pool_size = 10 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.141871] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] vmware.console_delay_seconds = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.142050] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] vmware.datastore_regex = ^datastore.* {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.142254] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] vmware.host_ip = vc1.osci.c.eu-de-1.cloud.sap {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.142427] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] vmware.host_password = **** {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.142594] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] vmware.host_port = 443 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.142764] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] vmware.host_username = administrator@vsphere.local {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.142931] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] vmware.insecure = True {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.143105] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] vmware.integration_bridge = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.143275] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] vmware.maximum_objects = 100 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.143434] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] vmware.pbm_default_policy = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.143596] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] vmware.pbm_enabled = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.143757] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] vmware.pbm_wsdl_location = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.143927] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] vmware.serial_log_dir = /opt/vmware/vspc {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.144098] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] vmware.serial_port_proxy_uri = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.144261] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] vmware.serial_port_service_uri = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.144426] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] vmware.task_poll_interval = 0.5 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.144598] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] vmware.use_linked_clone = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.144769] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] vmware.vnc_keymap = en-us {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.144934] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] vmware.vnc_port = 5900 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.145114] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] vmware.vnc_port_total = 10000 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.145300] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] vnc.auth_schemes = ['none'] {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.145474] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] vnc.enabled = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.145769] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] vnc.novncproxy_base_url = http://127.0.0.1:6080/vnc_auto.html {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.145964] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] vnc.novncproxy_host = 0.0.0.0 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.146162] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] vnc.novncproxy_port = 6080 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.146340] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] vnc.server_listen = 127.0.0.1 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.146537] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] vnc.server_proxyclient_address = 127.0.0.1 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.146709] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] vnc.vencrypt_ca_certs = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.146871] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] vnc.vencrypt_client_cert = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.147040] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] vnc.vencrypt_client_key = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.147227] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] workarounds.disable_compute_service_check_for_ffu = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.147391] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] workarounds.disable_deep_image_inspection = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.147573] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] workarounds.disable_fallback_pcpu_query = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.147746] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] workarounds.disable_group_policy_check_upcall = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.147910] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] workarounds.disable_libvirt_livesnapshot = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.148086] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] workarounds.disable_rootwrap = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.148253] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] workarounds.enable_numa_live_migration = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.148413] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] workarounds.enable_qemu_monitor_announce_self = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.148575] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] workarounds.ensure_libvirt_rbd_instance_dir_cleanup = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.148738] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] workarounds.handle_virt_lifecycle_events = True {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.148931] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] workarounds.libvirt_disable_apic = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.149118] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] workarounds.never_download_image_if_on_rbd = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.149288] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] workarounds.qemu_monitor_announce_self_count = 3 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.149454] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] workarounds.qemu_monitor_announce_self_interval = 1 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.149617] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] workarounds.reserve_disk_resource_for_image_cache = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.149780] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] workarounds.skip_cpu_compare_at_startup = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.149943] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] workarounds.skip_cpu_compare_on_dest = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.150117] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] workarounds.skip_hypervisor_version_check_on_lm = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.150282] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] workarounds.skip_reserve_in_use_ironic_nodes = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.150442] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] workarounds.unified_limits_count_pcpu_as_vcpu = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.150607] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] workarounds.wait_for_vif_plugged_event_during_hard_reboot = [] {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.150792] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] wsgi.api_paste_config = /etc/nova/api-paste.ini {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.150962] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] wsgi.client_socket_timeout = 900 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.151144] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] wsgi.default_pool_size = 1000 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.151310] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] wsgi.keep_alive = True {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.151479] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] wsgi.max_header_line = 16384 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.151687] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] wsgi.secure_proxy_ssl_header = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.151874] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] wsgi.ssl_ca_file = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.152051] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] wsgi.ssl_cert_file = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.152220] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] wsgi.ssl_key_file = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.152387] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] wsgi.tcp_keepidle = 600 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.152568] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] wsgi.wsgi_log_format = %(client_ip)s "%(request_line)s" status: %(status_code)s len: %(body_length)s time: %(wall_seconds).7f {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.152736] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] zvm.ca_file = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.152897] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] zvm.cloud_connector_url = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.153192] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] zvm.image_tmp_path = /opt/stack/data/n-cpu-1/images {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.153372] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] zvm.reachable_timeout = 300 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.153553] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_policy.enforce_new_defaults = True {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.153993] env[63088]: WARNING oslo_config.cfg [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] Deprecated: Option "enforce_scope" from group "oslo_policy" is deprecated for removal (This configuration was added temporarily to facilitate a smooth transition to the new RBAC. OpenStack will always enforce scope checks. This configuration option is deprecated and will be removed in the 2025.2 cycle.). Its value may be silently ignored in the future. [ 497.154211] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_policy.enforce_scope = True {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.154401] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_policy.policy_default_rule = default {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.154585] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_policy.policy_dirs = ['policy.d'] {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.154764] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_policy.policy_file = policy.yaml {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.154946] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_policy.remote_content_type = application/x-www-form-urlencoded {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.155124] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_policy.remote_ssl_ca_crt_file = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.155295] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_policy.remote_ssl_client_crt_file = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.155452] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_policy.remote_ssl_client_key_file = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.155616] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_policy.remote_ssl_verify_server_crt = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.155785] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_versionedobjects.fatal_exception_format_errors = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.155962] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_middleware.http_basic_auth_user_file = /etc/htpasswd {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.156154] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] profiler.connection_string = messaging:// {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.156325] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] profiler.enabled = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.156532] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] profiler.es_doc_type = notification {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.156691] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] profiler.es_scroll_size = 10000 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.156866] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] profiler.es_scroll_time = 2m {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.157043] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] profiler.filter_error_trace = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.157282] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] profiler.hmac_keys = **** {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.157498] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] profiler.sentinel_service_name = mymaster {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.157695] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] profiler.socket_timeout = 0.1 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.157869] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] profiler.trace_requests = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.158046] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] profiler.trace_sqlalchemy = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.158238] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] profiler_jaeger.process_tags = {} {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.158403] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] profiler_jaeger.service_name_prefix = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.158571] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] profiler_otlp.service_name_prefix = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.158738] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] remote_debug.host = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.158900] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] remote_debug.port = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.159098] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_messaging_rabbit.amqp_auto_delete = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.159268] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_messaging_rabbit.amqp_durable_queues = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.159436] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_messaging_rabbit.conn_pool_min_size = 2 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.159601] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_messaging_rabbit.conn_pool_ttl = 1200 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.159767] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_messaging_rabbit.direct_mandatory_flag = True {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.159931] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_messaging_rabbit.enable_cancel_on_failover = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.160105] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_messaging_rabbit.heartbeat_in_pthread = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.160273] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_messaging_rabbit.heartbeat_rate = 3 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.160436] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.160607] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_messaging_rabbit.hostname = devstack {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.160771] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_messaging_rabbit.kombu_compression = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.160944] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_messaging_rabbit.kombu_failover_strategy = round-robin {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.161126] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.161299] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.161471] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_messaging_rabbit.processname = nova-compute {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.161648] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_messaging_rabbit.rabbit_ha_queues = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.161891] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_messaging_rabbit.rabbit_interval_max = 30 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.162103] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.162280] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 0 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.162447] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.162617] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.162787] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.162950] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_messaging_rabbit.rabbit_quorum_queue = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.163134] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_messaging_rabbit.rabbit_retry_backoff = 2 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.163302] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_messaging_rabbit.rabbit_retry_interval = 1 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.163465] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_messaging_rabbit.rabbit_stream_fanout = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.163628] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.163791] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_messaging_rabbit.rabbit_transient_quorum_queue = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.163959] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_messaging_rabbit.rpc_conn_pool_size = 30 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.164140] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_messaging_rabbit.ssl = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.164315] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_messaging_rabbit.ssl_ca_file = {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.164487] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_messaging_rabbit.ssl_cert_file = {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.164652] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_messaging_rabbit.ssl_enforce_fips_mode = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.164822] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_messaging_rabbit.ssl_key_file = {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.164991] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_messaging_rabbit.ssl_version = {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.165169] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_messaging_rabbit.use_queue_manager = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.165389] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_messaging_notifications.driver = ['messagingv2'] {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.165524] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_messaging_notifications.retry = -1 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.165710] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_messaging_notifications.topics = ['notifications'] {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.165889] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_messaging_notifications.transport_url = **** {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.166070] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_limit.auth_section = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.166240] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_limit.auth_type = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.166401] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_limit.cafile = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.166593] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_limit.certfile = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.166763] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_limit.collect_timing = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.166931] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_limit.connect_retries = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.167098] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_limit.connect_retry_delay = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.167257] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_limit.endpoint_id = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.167418] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_limit.endpoint_override = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.167631] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_limit.insecure = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.167816] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_limit.keyfile = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.167979] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_limit.max_version = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.168154] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_limit.min_version = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.168315] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_limit.region_name = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.168478] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_limit.retriable_status_codes = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.168636] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_limit.service_name = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.168816] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_limit.service_type = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.169053] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_limit.split_loggers = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.169252] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_limit.status_code_retries = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.169541] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_limit.status_code_retry_delay = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.169751] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_limit.timeout = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.169925] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_limit.valid_interfaces = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.170102] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_limit.version = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.170275] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_reports.file_event_handler = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.170445] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_reports.file_event_handler_interval = 1 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.170604] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] oslo_reports.log_dir = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.170871] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] vif_plug_linux_bridge_privileged.capabilities = [12] {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.171182] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] vif_plug_linux_bridge_privileged.group = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.171425] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] vif_plug_linux_bridge_privileged.helper_command = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.171623] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] vif_plug_linux_bridge_privileged.logger_name = oslo_privsep.daemon {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.171800] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] vif_plug_linux_bridge_privileged.thread_pool_size = 8 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.171964] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] vif_plug_linux_bridge_privileged.user = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.172156] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] vif_plug_ovs_privileged.capabilities = [12, 1] {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.172319] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] vif_plug_ovs_privileged.group = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.172479] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] vif_plug_ovs_privileged.helper_command = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.172645] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] vif_plug_ovs_privileged.logger_name = oslo_privsep.daemon {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.172812] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] vif_plug_ovs_privileged.thread_pool_size = 8 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.173018] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] vif_plug_ovs_privileged.user = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.173270] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] os_vif_linux_bridge.flat_interface = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.173549] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] os_vif_linux_bridge.forward_bridge_interface = ['all'] {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.173767] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] os_vif_linux_bridge.iptables_bottom_regex = {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.173957] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] os_vif_linux_bridge.iptables_drop_action = DROP {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.174148] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] os_vif_linux_bridge.iptables_top_regex = {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.174321] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] os_vif_linux_bridge.network_device_mtu = 1500 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.174491] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] os_vif_linux_bridge.use_ipv6 = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.174657] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] os_vif_linux_bridge.vlan_interface = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.174843] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] os_vif_ovs.default_qos_type = linux-noop {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.175019] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] os_vif_ovs.isolate_vif = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.175200] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] os_vif_ovs.network_device_mtu = 1500 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.175370] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] os_vif_ovs.ovs_vsctl_timeout = 120 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.175543] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] os_vif_ovs.ovsdb_connection = tcp:127.0.0.1:6640 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.175715] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] os_vif_ovs.ovsdb_interface = native {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.175880] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] os_vif_ovs.per_port_bridge = False {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.176055] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] os_brick.lock_path = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.176227] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] os_brick.wait_mpath_device_attempts = 4 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.176390] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] os_brick.wait_mpath_device_interval = 1 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.176589] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] privsep_osbrick.capabilities = [21] {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.176767] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] privsep_osbrick.group = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.176925] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] privsep_osbrick.helper_command = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.177104] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] privsep_osbrick.logger_name = os_brick.privileged {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.177271] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] privsep_osbrick.thread_pool_size = 8 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.177430] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] privsep_osbrick.user = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.177700] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] nova_sys_admin.capabilities = [0, 1, 2, 3, 12, 21] {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.177910] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] nova_sys_admin.group = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.178097] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] nova_sys_admin.helper_command = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.178272] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] nova_sys_admin.logger_name = oslo_privsep.daemon {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.178439] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] nova_sys_admin.thread_pool_size = 8 {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.178602] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] nova_sys_admin.user = None {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 497.178737] env[63088]: DEBUG oslo_service.service [None req-9d564c6a-80fd-4642-88a2-14e40dd2af47 None None] ******************************************************************************** {{(pid=63088) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2830}} [ 497.179276] env[63088]: INFO nova.service [-] Starting compute node (version 0.0.1) [ 497.682821] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-4f4a2fa2-c6a1-42a4-8c8b-878a7ce410f7 None None] Getting list of instances from cluster (obj){ [ 497.682821] env[63088]: value = "domain-c8" [ 497.682821] env[63088]: _type = "ClusterComputeResource" [ 497.682821] env[63088]: } {{(pid=63088) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 497.684157] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e31e4419-bccb-4fd0-94a0-13388193330c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 497.694182] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-4f4a2fa2-c6a1-42a4-8c8b-878a7ce410f7 None None] Got total of 0 instances {{(pid=63088) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 497.694182] env[63088]: WARNING nova.virt.vmwareapi.driver [None req-4f4a2fa2-c6a1-42a4-8c8b-878a7ce410f7 None None] The vmwareapi driver is not tested by the OpenStack project nor does it have clear maintainer(s) and thus its quality can not be ensured. It should be considered experimental and may be removed in a future release. If you are using the driver in production please let us know via the openstack-discuss mailing list. [ 497.694353] env[63088]: INFO nova.virt.node [None req-4f4a2fa2-c6a1-42a4-8c8b-878a7ce410f7 None None] Generated node identity 6eae54a9-8831-40eb-bf54-4bc60d346b02 [ 497.694549] env[63088]: INFO nova.virt.node [None req-4f4a2fa2-c6a1-42a4-8c8b-878a7ce410f7 None None] Wrote node identity 6eae54a9-8831-40eb-bf54-4bc60d346b02 to /opt/stack/data/n-cpu-1/compute_id [ 498.197109] env[63088]: WARNING nova.compute.manager [None req-4f4a2fa2-c6a1-42a4-8c8b-878a7ce410f7 None None] Compute nodes ['6eae54a9-8831-40eb-bf54-4bc60d346b02'] for host cpu-1 were not found in the database. If this is the first time this service is starting on this host, then you can ignore this warning. [ 499.203650] env[63088]: INFO nova.compute.manager [None req-4f4a2fa2-c6a1-42a4-8c8b-878a7ce410f7 None None] Looking for unclaimed instances stuck in BUILDING status for nodes managed by this host [ 500.209089] env[63088]: WARNING nova.compute.manager [None req-4f4a2fa2-c6a1-42a4-8c8b-878a7ce410f7 None None] No compute node record found for host cpu-1. If this is the first time this service is starting on this host, then you can ignore this warning.: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host cpu-1 could not be found. [ 500.209451] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4f4a2fa2-c6a1-42a4-8c8b-878a7ce410f7 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 500.209565] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4f4a2fa2-c6a1-42a4-8c8b-878a7ce410f7 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 500.209703] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4f4a2fa2-c6a1-42a4-8c8b-878a7ce410f7 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 500.209860] env[63088]: DEBUG nova.compute.resource_tracker [None req-4f4a2fa2-c6a1-42a4-8c8b-878a7ce410f7 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63088) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 500.210834] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db0af107-099c-4fb6-bc3e-f7883204fab9 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 500.219635] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-177abda9-09fe-4021-b8ae-faa9f368cba3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 500.235615] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d79cdd7d-3413-4929-8b62-d552e2444286 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 500.242241] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3f199b2-416d-43ae-bb84-c78f8971c6c6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 500.272067] env[63088]: DEBUG nova.compute.resource_tracker [None req-4f4a2fa2-c6a1-42a4-8c8b-878a7ce410f7 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181497MB free_disk=141GB free_vcpus=48 pci_devices=None {{(pid=63088) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 500.272227] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4f4a2fa2-c6a1-42a4-8c8b-878a7ce410f7 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 500.272424] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4f4a2fa2-c6a1-42a4-8c8b-878a7ce410f7 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 500.775220] env[63088]: WARNING nova.compute.resource_tracker [None req-4f4a2fa2-c6a1-42a4-8c8b-878a7ce410f7 None None] No compute node record for cpu-1:6eae54a9-8831-40eb-bf54-4bc60d346b02: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host 6eae54a9-8831-40eb-bf54-4bc60d346b02 could not be found. [ 501.278934] env[63088]: INFO nova.compute.resource_tracker [None req-4f4a2fa2-c6a1-42a4-8c8b-878a7ce410f7 None None] Compute node record created for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 with uuid: 6eae54a9-8831-40eb-bf54-4bc60d346b02 [ 502.787522] env[63088]: DEBUG nova.compute.resource_tracker [None req-4f4a2fa2-c6a1-42a4-8c8b-878a7ce410f7 None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=63088) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 502.788035] env[63088]: DEBUG nova.compute.resource_tracker [None req-4f4a2fa2-c6a1-42a4-8c8b-878a7ce410f7 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=63088) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 502.941450] env[63088]: INFO nova.scheduler.client.report [None req-4f4a2fa2-c6a1-42a4-8c8b-878a7ce410f7 None None] [req-1dd9205e-16d2-47eb-9b04-ff28f08423bc] Created resource provider record via placement API for resource provider with UUID 6eae54a9-8831-40eb-bf54-4bc60d346b02 and name domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28. [ 502.956880] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91a9768f-7eb1-4e46-ae9c-3538de2dc257 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 502.964747] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-634f6da4-5020-43d4-a5c9-f5025172f16b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 502.993868] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80273957-f987-497b-a406-5dd188ab4157 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 503.000543] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93f4f643-e8c6-463b-a201-df77053ef27e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 503.012836] env[63088]: DEBUG nova.compute.provider_tree [None req-4f4a2fa2-c6a1-42a4-8c8b-878a7ce410f7 None None] Updating inventory in ProviderTree for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 503.548814] env[63088]: DEBUG nova.scheduler.client.report [None req-4f4a2fa2-c6a1-42a4-8c8b-878a7ce410f7 None None] Updated inventory for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 with generation 0 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 503.549225] env[63088]: DEBUG nova.compute.provider_tree [None req-4f4a2fa2-c6a1-42a4-8c8b-878a7ce410f7 None None] Updating resource provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 generation from 0 to 1 during operation: update_inventory {{(pid=63088) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 503.549399] env[63088]: DEBUG nova.compute.provider_tree [None req-4f4a2fa2-c6a1-42a4-8c8b-878a7ce410f7 None None] Updating inventory in ProviderTree for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 503.600861] env[63088]: DEBUG nova.compute.provider_tree [None req-4f4a2fa2-c6a1-42a4-8c8b-878a7ce410f7 None None] Updating resource provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 generation from 1 to 2 during operation: update_traits {{(pid=63088) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 504.105442] env[63088]: DEBUG nova.compute.resource_tracker [None req-4f4a2fa2-c6a1-42a4-8c8b-878a7ce410f7 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63088) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 504.105807] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4f4a2fa2-c6a1-42a4-8c8b-878a7ce410f7 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.833s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 504.105807] env[63088]: DEBUG nova.service [None req-4f4a2fa2-c6a1-42a4-8c8b-878a7ce410f7 None None] Creating RPC server for service compute {{(pid=63088) start /opt/stack/nova/nova/service.py:186}} [ 504.119480] env[63088]: DEBUG nova.service [None req-4f4a2fa2-c6a1-42a4-8c8b-878a7ce410f7 None None] Join ServiceGroup membership for this service compute {{(pid=63088) start /opt/stack/nova/nova/service.py:203}} [ 504.119650] env[63088]: DEBUG nova.servicegroup.drivers.db [None req-4f4a2fa2-c6a1-42a4-8c8b-878a7ce410f7 None None] DB_Driver: join new ServiceGroup member cpu-1 to the compute group, service = {{(pid=63088) join /opt/stack/nova/nova/servicegroup/drivers/db.py:44}} [ 532.124811] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._sync_power_states {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 532.627901] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Getting list of instances from cluster (obj){ [ 532.627901] env[63088]: value = "domain-c8" [ 532.627901] env[63088]: _type = "ClusterComputeResource" [ 532.627901] env[63088]: } {{(pid=63088) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 532.629124] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe2d9f14-98d3-46fe-a4c3-69b480d962eb {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.639382] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Got total of 0 instances {{(pid=63088) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 532.639622] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._cleanup_running_deleted_instances {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 532.639933] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Getting list of instances from cluster (obj){ [ 532.639933] env[63088]: value = "domain-c8" [ 532.639933] env[63088]: _type = "ClusterComputeResource" [ 532.639933] env[63088]: } {{(pid=63088) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 532.641217] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5ab388b-7b48-477e-bd5c-d13904e6bb02 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.649351] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Got total of 0 instances {{(pid=63088) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 543.783890] env[63088]: DEBUG oslo_concurrency.lockutils [None req-96b6b7c1-255c-445f-b8db-9f4849d061fb tempest-ImagesNegativeTestJSON-1434405362 tempest-ImagesNegativeTestJSON-1434405362-project-member] Acquiring lock "0fd65b6b-2363-4b45-bc07-365157a8b052" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 543.784203] env[63088]: DEBUG oslo_concurrency.lockutils [None req-96b6b7c1-255c-445f-b8db-9f4849d061fb tempest-ImagesNegativeTestJSON-1434405362 tempest-ImagesNegativeTestJSON-1434405362-project-member] Lock "0fd65b6b-2363-4b45-bc07-365157a8b052" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 544.215081] env[63088]: DEBUG oslo_concurrency.lockutils [None req-fa54bfad-a3c9-4ca3-8f38-8bc8499d1488 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] Acquiring lock "05d2ad93-949e-42e3-85e9-eed9baaebe8c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 544.215825] env[63088]: DEBUG oslo_concurrency.lockutils [None req-fa54bfad-a3c9-4ca3-8f38-8bc8499d1488 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] Lock "05d2ad93-949e-42e3-85e9-eed9baaebe8c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 544.286654] env[63088]: DEBUG nova.compute.manager [None req-96b6b7c1-255c-445f-b8db-9f4849d061fb tempest-ImagesNegativeTestJSON-1434405362 tempest-ImagesNegativeTestJSON-1434405362-project-member] [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 544.509822] env[63088]: DEBUG oslo_concurrency.lockutils [None req-710306d6-422f-4cdd-91a1-33da212e5f03 tempest-ServersAdminNegativeTestJSON-229078574 tempest-ServersAdminNegativeTestJSON-229078574-project-member] Acquiring lock "4f487d47-daf7-4ef1-9e8b-8d8f870bfe52" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 544.511087] env[63088]: DEBUG oslo_concurrency.lockutils [None req-710306d6-422f-4cdd-91a1-33da212e5f03 tempest-ServersAdminNegativeTestJSON-229078574 tempest-ServersAdminNegativeTestJSON-229078574-project-member] Lock "4f487d47-daf7-4ef1-9e8b-8d8f870bfe52" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 544.717586] env[63088]: DEBUG nova.compute.manager [None req-fa54bfad-a3c9-4ca3-8f38-8bc8499d1488 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 544.758723] env[63088]: DEBUG oslo_concurrency.lockutils [None req-8a34ea86-0280-447d-8082-40a2090a8af5 tempest-ServerDiagnosticsTest-1360445823 tempest-ServerDiagnosticsTest-1360445823-project-member] Acquiring lock "2d7bf024-afb5-4d33-9c5e-06633167ac77" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 544.758978] env[63088]: DEBUG oslo_concurrency.lockutils [None req-8a34ea86-0280-447d-8082-40a2090a8af5 tempest-ServerDiagnosticsTest-1360445823 tempest-ServerDiagnosticsTest-1360445823-project-member] Lock "2d7bf024-afb5-4d33-9c5e-06633167ac77" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 544.835830] env[63088]: DEBUG oslo_concurrency.lockutils [None req-96b6b7c1-255c-445f-b8db-9f4849d061fb tempest-ImagesNegativeTestJSON-1434405362 tempest-ImagesNegativeTestJSON-1434405362-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 544.836096] env[63088]: DEBUG oslo_concurrency.lockutils [None req-96b6b7c1-255c-445f-b8db-9f4849d061fb tempest-ImagesNegativeTestJSON-1434405362 tempest-ImagesNegativeTestJSON-1434405362-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 544.838136] env[63088]: INFO nova.compute.claims [None req-96b6b7c1-255c-445f-b8db-9f4849d061fb tempest-ImagesNegativeTestJSON-1434405362 tempest-ImagesNegativeTestJSON-1434405362-project-member] [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 545.012565] env[63088]: DEBUG nova.compute.manager [None req-710306d6-422f-4cdd-91a1-33da212e5f03 tempest-ServersAdminNegativeTestJSON-229078574 tempest-ServersAdminNegativeTestJSON-229078574-project-member] [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 545.247547] env[63088]: DEBUG oslo_concurrency.lockutils [None req-fa54bfad-a3c9-4ca3-8f38-8bc8499d1488 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 545.263803] env[63088]: DEBUG nova.compute.manager [None req-8a34ea86-0280-447d-8082-40a2090a8af5 tempest-ServerDiagnosticsTest-1360445823 tempest-ServerDiagnosticsTest-1360445823-project-member] [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 545.541493] env[63088]: DEBUG oslo_concurrency.lockutils [None req-710306d6-422f-4cdd-91a1-33da212e5f03 tempest-ServersAdminNegativeTestJSON-229078574 tempest-ServersAdminNegativeTestJSON-229078574-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 545.796411] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Acquiring lock "e8cfbce4-cfb2-463e-9f38-78a058dab6e1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 545.796673] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Lock "e8cfbce4-cfb2-463e-9f38-78a058dab6e1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 545.799578] env[63088]: DEBUG oslo_concurrency.lockutils [None req-8a34ea86-0280-447d-8082-40a2090a8af5 tempest-ServerDiagnosticsTest-1360445823 tempest-ServerDiagnosticsTest-1360445823-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 545.975532] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5040527-a6e2-4efa-ba9c-424d8b5fd29a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.987837] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-212c9d13-48e3-4da5-ade8-0713c1ca8392 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.032919] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b98be864-eefa-4484-8dfb-3098714549c6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.042230] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77bd1e4a-4b68-47be-beeb-070b7c5ee7ca {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.058916] env[63088]: DEBUG nova.compute.provider_tree [None req-96b6b7c1-255c-445f-b8db-9f4849d061fb tempest-ImagesNegativeTestJSON-1434405362 tempest-ImagesNegativeTestJSON-1434405362-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 546.299549] env[63088]: DEBUG nova.compute.manager [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 546.559468] env[63088]: DEBUG oslo_concurrency.lockutils [None req-61dd383c-8cef-4a29-a464-69dfd03cb5ce tempest-ServerExternalEventsTest-515168927 tempest-ServerExternalEventsTest-515168927-project-member] Acquiring lock "173fccf1-f258-4f0c-a968-466a27406c55" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 546.559842] env[63088]: DEBUG oslo_concurrency.lockutils [None req-61dd383c-8cef-4a29-a464-69dfd03cb5ce tempest-ServerExternalEventsTest-515168927 tempest-ServerExternalEventsTest-515168927-project-member] Lock "173fccf1-f258-4f0c-a968-466a27406c55" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 546.566425] env[63088]: DEBUG nova.scheduler.client.report [None req-96b6b7c1-255c-445f-b8db-9f4849d061fb tempest-ImagesNegativeTestJSON-1434405362 tempest-ImagesNegativeTestJSON-1434405362-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 546.825844] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 547.064070] env[63088]: DEBUG nova.compute.manager [None req-61dd383c-8cef-4a29-a464-69dfd03cb5ce tempest-ServerExternalEventsTest-515168927 tempest-ServerExternalEventsTest-515168927-project-member] [instance: 173fccf1-f258-4f0c-a968-466a27406c55] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 547.071142] env[63088]: DEBUG oslo_concurrency.lockutils [None req-96b6b7c1-255c-445f-b8db-9f4849d061fb tempest-ImagesNegativeTestJSON-1434405362 tempest-ImagesNegativeTestJSON-1434405362-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.235s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 547.071755] env[63088]: DEBUG nova.compute.manager [None req-96b6b7c1-255c-445f-b8db-9f4849d061fb tempest-ImagesNegativeTestJSON-1434405362 tempest-ImagesNegativeTestJSON-1434405362-project-member] [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 547.080275] env[63088]: DEBUG oslo_concurrency.lockutils [None req-fa54bfad-a3c9-4ca3-8f38-8bc8499d1488 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.831s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 547.080275] env[63088]: INFO nova.compute.claims [None req-fa54bfad-a3c9-4ca3-8f38-8bc8499d1488 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 547.590031] env[63088]: DEBUG nova.compute.utils [None req-96b6b7c1-255c-445f-b8db-9f4849d061fb tempest-ImagesNegativeTestJSON-1434405362 tempest-ImagesNegativeTestJSON-1434405362-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 547.590401] env[63088]: DEBUG nova.compute.manager [None req-96b6b7c1-255c-445f-b8db-9f4849d061fb tempest-ImagesNegativeTestJSON-1434405362 tempest-ImagesNegativeTestJSON-1434405362-project-member] [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 547.590909] env[63088]: DEBUG nova.network.neutron [None req-96b6b7c1-255c-445f-b8db-9f4849d061fb tempest-ImagesNegativeTestJSON-1434405362 tempest-ImagesNegativeTestJSON-1434405362-project-member] [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 547.599022] env[63088]: DEBUG oslo_concurrency.lockutils [None req-61dd383c-8cef-4a29-a464-69dfd03cb5ce tempest-ServerExternalEventsTest-515168927 tempest-ServerExternalEventsTest-515168927-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 547.641518] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bc610745-9520-42a6-9fd1-7e4c9c63b77a tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] Acquiring lock "5c3f72e3-d4f4-4c34-92a3-075536b85c99" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 547.641725] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bc610745-9520-42a6-9fd1-7e4c9c63b77a tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] Lock "5c3f72e3-d4f4-4c34-92a3-075536b85c99" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 547.925364] env[63088]: DEBUG nova.policy [None req-96b6b7c1-255c-445f-b8db-9f4849d061fb tempest-ImagesNegativeTestJSON-1434405362 tempest-ImagesNegativeTestJSON-1434405362-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ffbd855b12994dcc8c46809db4d1994a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '04db3173c8924b02bf3277cb0be1db84', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 548.100577] env[63088]: DEBUG nova.compute.manager [None req-96b6b7c1-255c-445f-b8db-9f4849d061fb tempest-ImagesNegativeTestJSON-1434405362 tempest-ImagesNegativeTestJSON-1434405362-project-member] [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 548.147530] env[63088]: DEBUG nova.compute.manager [None req-bc610745-9520-42a6-9fd1-7e4c9c63b77a tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 548.274496] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-155f19da-8947-4622-bbcc-e695c33ce425 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.284623] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ebd7f0a-6925-4dd0-8ce2-821a0e392644 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.289433] env[63088]: DEBUG nova.network.neutron [None req-96b6b7c1-255c-445f-b8db-9f4849d061fb tempest-ImagesNegativeTestJSON-1434405362 tempest-ImagesNegativeTestJSON-1434405362-project-member] [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] Successfully created port: 946407e1-ac24-4d57-9f2f-f2c0994ff966 {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 548.318216] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55ce67f3-74be-48b1-86e7-88ac7175a46c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.326086] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b05f4794-d18d-4bac-b24c-38ca1eea9c77 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.346286] env[63088]: DEBUG nova.compute.provider_tree [None req-fa54bfad-a3c9-4ca3-8f38-8bc8499d1488 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 548.684345] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bc610745-9520-42a6-9fd1-7e4c9c63b77a tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 548.849657] env[63088]: DEBUG nova.scheduler.client.report [None req-fa54bfad-a3c9-4ca3-8f38-8bc8499d1488 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 549.114902] env[63088]: DEBUG nova.compute.manager [None req-96b6b7c1-255c-445f-b8db-9f4849d061fb tempest-ImagesNegativeTestJSON-1434405362 tempest-ImagesNegativeTestJSON-1434405362-project-member] [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 549.157494] env[63088]: DEBUG nova.virt.hardware [None req-96b6b7c1-255c-445f-b8db-9f4849d061fb tempest-ImagesNegativeTestJSON-1434405362 tempest-ImagesNegativeTestJSON-1434405362-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 549.157786] env[63088]: DEBUG nova.virt.hardware [None req-96b6b7c1-255c-445f-b8db-9f4849d061fb tempest-ImagesNegativeTestJSON-1434405362 tempest-ImagesNegativeTestJSON-1434405362-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 549.157935] env[63088]: DEBUG nova.virt.hardware [None req-96b6b7c1-255c-445f-b8db-9f4849d061fb tempest-ImagesNegativeTestJSON-1434405362 tempest-ImagesNegativeTestJSON-1434405362-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 549.158763] env[63088]: DEBUG nova.virt.hardware [None req-96b6b7c1-255c-445f-b8db-9f4849d061fb tempest-ImagesNegativeTestJSON-1434405362 tempest-ImagesNegativeTestJSON-1434405362-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 549.158763] env[63088]: DEBUG nova.virt.hardware [None req-96b6b7c1-255c-445f-b8db-9f4849d061fb tempest-ImagesNegativeTestJSON-1434405362 tempest-ImagesNegativeTestJSON-1434405362-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 549.158763] env[63088]: DEBUG nova.virt.hardware [None req-96b6b7c1-255c-445f-b8db-9f4849d061fb tempest-ImagesNegativeTestJSON-1434405362 tempest-ImagesNegativeTestJSON-1434405362-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 549.158763] env[63088]: DEBUG nova.virt.hardware [None req-96b6b7c1-255c-445f-b8db-9f4849d061fb tempest-ImagesNegativeTestJSON-1434405362 tempest-ImagesNegativeTestJSON-1434405362-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 549.158763] env[63088]: DEBUG nova.virt.hardware [None req-96b6b7c1-255c-445f-b8db-9f4849d061fb tempest-ImagesNegativeTestJSON-1434405362 tempest-ImagesNegativeTestJSON-1434405362-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 549.160083] env[63088]: DEBUG nova.virt.hardware [None req-96b6b7c1-255c-445f-b8db-9f4849d061fb tempest-ImagesNegativeTestJSON-1434405362 tempest-ImagesNegativeTestJSON-1434405362-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 549.160083] env[63088]: DEBUG nova.virt.hardware [None req-96b6b7c1-255c-445f-b8db-9f4849d061fb tempest-ImagesNegativeTestJSON-1434405362 tempest-ImagesNegativeTestJSON-1434405362-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 549.160083] env[63088]: DEBUG nova.virt.hardware [None req-96b6b7c1-255c-445f-b8db-9f4849d061fb tempest-ImagesNegativeTestJSON-1434405362 tempest-ImagesNegativeTestJSON-1434405362-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 549.160712] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac6bcc0f-688f-476f-92e9-2bd8611bcf43 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.171966] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6568dcbe-ebf7-46f4-8cbf-a6cec6f2c264 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] Acquiring lock "4d5293a0-6945-4731-afb6-996a142c8447" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 549.172217] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6568dcbe-ebf7-46f4-8cbf-a6cec6f2c264 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] Lock "4d5293a0-6945-4731-afb6-996a142c8447" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 549.174452] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e298e8c-6661-41f9-a89e-5c23446bb309 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.195625] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07f5ead0-a667-414e-ad4e-7d8b41be4bf5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.356160] env[63088]: DEBUG oslo_concurrency.lockutils [None req-fa54bfad-a3c9-4ca3-8f38-8bc8499d1488 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.279s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 549.356925] env[63088]: DEBUG nova.compute.manager [None req-fa54bfad-a3c9-4ca3-8f38-8bc8499d1488 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 549.360148] env[63088]: DEBUG oslo_concurrency.lockutils [None req-710306d6-422f-4cdd-91a1-33da212e5f03 tempest-ServersAdminNegativeTestJSON-229078574 tempest-ServersAdminNegativeTestJSON-229078574-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.819s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 549.362739] env[63088]: INFO nova.compute.claims [None req-710306d6-422f-4cdd-91a1-33da212e5f03 tempest-ServersAdminNegativeTestJSON-229078574 tempest-ServersAdminNegativeTestJSON-229078574-project-member] [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 549.679340] env[63088]: DEBUG nova.compute.manager [None req-6568dcbe-ebf7-46f4-8cbf-a6cec6f2c264 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] [instance: 4d5293a0-6945-4731-afb6-996a142c8447] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 549.871729] env[63088]: DEBUG nova.compute.utils [None req-fa54bfad-a3c9-4ca3-8f38-8bc8499d1488 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 549.878164] env[63088]: DEBUG nova.compute.manager [None req-fa54bfad-a3c9-4ca3-8f38-8bc8499d1488 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 549.878477] env[63088]: DEBUG nova.network.neutron [None req-fa54bfad-a3c9-4ca3-8f38-8bc8499d1488 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 550.080571] env[63088]: DEBUG nova.policy [None req-fa54bfad-a3c9-4ca3-8f38-8bc8499d1488 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ba9caaadce804132b4df2bc441099f06', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd593db530697467cad3671cc591a117e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 550.217243] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6568dcbe-ebf7-46f4-8cbf-a6cec6f2c264 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 550.378752] env[63088]: DEBUG nova.compute.manager [None req-fa54bfad-a3c9-4ca3-8f38-8bc8499d1488 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 550.563254] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f40d6d5-7272-4b24-9a69-8c11cfa32d1a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.572251] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-142a63e4-09c9-439d-926d-329e14f0973c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.605502] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c95fa5e0-62f6-44ae-8dd0-f3480b1d5584 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.615541] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcd8d9eb-f736-415e-805d-53f922cf6834 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.630295] env[63088]: DEBUG nova.compute.provider_tree [None req-710306d6-422f-4cdd-91a1-33da212e5f03 tempest-ServersAdminNegativeTestJSON-229078574 tempest-ServersAdminNegativeTestJSON-229078574-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 551.090532] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Acquiring lock "c4777721-3f65-455f-9973-c1ed0732de34" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 551.090532] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Lock "c4777721-3f65-455f-9973-c1ed0732de34" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 551.133453] env[63088]: DEBUG nova.scheduler.client.report [None req-710306d6-422f-4cdd-91a1-33da212e5f03 tempest-ServersAdminNegativeTestJSON-229078574 tempest-ServersAdminNegativeTestJSON-229078574-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 551.398906] env[63088]: DEBUG nova.compute.manager [None req-fa54bfad-a3c9-4ca3-8f38-8bc8499d1488 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 551.413121] env[63088]: DEBUG nova.network.neutron [None req-fa54bfad-a3c9-4ca3-8f38-8bc8499d1488 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] Successfully created port: f66bce24-6dbd-4128-9ab8-86cf1c74d872 {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 551.436713] env[63088]: DEBUG nova.virt.hardware [None req-fa54bfad-a3c9-4ca3-8f38-8bc8499d1488 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 551.437536] env[63088]: DEBUG nova.virt.hardware [None req-fa54bfad-a3c9-4ca3-8f38-8bc8499d1488 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 551.437536] env[63088]: DEBUG nova.virt.hardware [None req-fa54bfad-a3c9-4ca3-8f38-8bc8499d1488 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 551.437536] env[63088]: DEBUG nova.virt.hardware [None req-fa54bfad-a3c9-4ca3-8f38-8bc8499d1488 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 551.437536] env[63088]: DEBUG nova.virt.hardware [None req-fa54bfad-a3c9-4ca3-8f38-8bc8499d1488 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 551.437752] env[63088]: DEBUG nova.virt.hardware [None req-fa54bfad-a3c9-4ca3-8f38-8bc8499d1488 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 551.437848] env[63088]: DEBUG nova.virt.hardware [None req-fa54bfad-a3c9-4ca3-8f38-8bc8499d1488 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 551.437981] env[63088]: DEBUG nova.virt.hardware [None req-fa54bfad-a3c9-4ca3-8f38-8bc8499d1488 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 551.439010] env[63088]: DEBUG nova.virt.hardware [None req-fa54bfad-a3c9-4ca3-8f38-8bc8499d1488 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 551.439220] env[63088]: DEBUG nova.virt.hardware [None req-fa54bfad-a3c9-4ca3-8f38-8bc8499d1488 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 551.439402] env[63088]: DEBUG nova.virt.hardware [None req-fa54bfad-a3c9-4ca3-8f38-8bc8499d1488 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 551.440344] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a77fddd4-77d3-48bb-bdda-04a40c9aa791 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.448835] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8997e2fb-ed08-4809-b5a1-3104636890e6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.593482] env[63088]: DEBUG nova.compute.manager [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] [instance: c4777721-3f65-455f-9973-c1ed0732de34] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 551.631986] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6e35ce22-cf37-483a-8093-9574de222716 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] Acquiring lock "44e39b2b-2bac-4a6e-9534-d2774d094931" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 551.632296] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6e35ce22-cf37-483a-8093-9574de222716 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] Lock "44e39b2b-2bac-4a6e-9534-d2774d094931" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 551.637899] env[63088]: DEBUG oslo_concurrency.lockutils [None req-710306d6-422f-4cdd-91a1-33da212e5f03 tempest-ServersAdminNegativeTestJSON-229078574 tempest-ServersAdminNegativeTestJSON-229078574-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.278s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 551.639616] env[63088]: DEBUG nova.compute.manager [None req-710306d6-422f-4cdd-91a1-33da212e5f03 tempest-ServersAdminNegativeTestJSON-229078574 tempest-ServersAdminNegativeTestJSON-229078574-project-member] [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 551.643053] env[63088]: DEBUG oslo_concurrency.lockutils [None req-8a34ea86-0280-447d-8082-40a2090a8af5 tempest-ServerDiagnosticsTest-1360445823 tempest-ServerDiagnosticsTest-1360445823-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.843s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 551.644151] env[63088]: INFO nova.compute.claims [None req-8a34ea86-0280-447d-8082-40a2090a8af5 tempest-ServerDiagnosticsTest-1360445823 tempest-ServerDiagnosticsTest-1360445823-project-member] [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 551.669342] env[63088]: ERROR nova.compute.manager [None req-96b6b7c1-255c-445f-b8db-9f4849d061fb tempest-ImagesNegativeTestJSON-1434405362 tempest-ImagesNegativeTestJSON-1434405362-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 946407e1-ac24-4d57-9f2f-f2c0994ff966, please check neutron logs for more information. [ 551.669342] env[63088]: ERROR nova.compute.manager Traceback (most recent call last): [ 551.669342] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 551.669342] env[63088]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 551.669342] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 551.669342] env[63088]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 551.669342] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 551.669342] env[63088]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 551.669342] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 551.669342] env[63088]: ERROR nova.compute.manager self.force_reraise() [ 551.669342] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 551.669342] env[63088]: ERROR nova.compute.manager raise self.value [ 551.669342] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 551.669342] env[63088]: ERROR nova.compute.manager updated_port = self._update_port( [ 551.669342] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 551.669342] env[63088]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 551.669829] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 551.669829] env[63088]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 551.669829] env[63088]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 946407e1-ac24-4d57-9f2f-f2c0994ff966, please check neutron logs for more information. [ 551.669829] env[63088]: ERROR nova.compute.manager [ 551.669829] env[63088]: Traceback (most recent call last): [ 551.669829] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 551.669829] env[63088]: listener.cb(fileno) [ 551.669829] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 551.669829] env[63088]: result = function(*args, **kwargs) [ 551.669829] env[63088]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 551.669829] env[63088]: return func(*args, **kwargs) [ 551.669829] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 551.669829] env[63088]: raise e [ 551.669829] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 551.669829] env[63088]: nwinfo = self.network_api.allocate_for_instance( [ 551.669829] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 551.669829] env[63088]: created_port_ids = self._update_ports_for_instance( [ 551.669829] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 551.669829] env[63088]: with excutils.save_and_reraise_exception(): [ 551.669829] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 551.669829] env[63088]: self.force_reraise() [ 551.669829] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 551.669829] env[63088]: raise self.value [ 551.669829] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 551.669829] env[63088]: updated_port = self._update_port( [ 551.669829] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 551.669829] env[63088]: _ensure_no_port_binding_failure(port) [ 551.669829] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 551.669829] env[63088]: raise exception.PortBindingFailed(port_id=port['id']) [ 551.670672] env[63088]: nova.exception.PortBindingFailed: Binding failed for port 946407e1-ac24-4d57-9f2f-f2c0994ff966, please check neutron logs for more information. [ 551.670672] env[63088]: Removing descriptor: 15 [ 551.675334] env[63088]: ERROR nova.compute.manager [None req-96b6b7c1-255c-445f-b8db-9f4849d061fb tempest-ImagesNegativeTestJSON-1434405362 tempest-ImagesNegativeTestJSON-1434405362-project-member] [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 946407e1-ac24-4d57-9f2f-f2c0994ff966, please check neutron logs for more information. [ 551.675334] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] Traceback (most recent call last): [ 551.675334] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 551.675334] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] yield resources [ 551.675334] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 551.675334] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] self.driver.spawn(context, instance, image_meta, [ 551.675334] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 551.675334] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] self._vmops.spawn(context, instance, image_meta, injected_files, [ 551.675334] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 551.675334] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] vm_ref = self.build_virtual_machine(instance, [ 551.675334] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 551.675660] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] vif_infos = vmwarevif.get_vif_info(self._session, [ 551.675660] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 551.675660] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] for vif in network_info: [ 551.675660] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 551.675660] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] return self._sync_wrapper(fn, *args, **kwargs) [ 551.675660] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 551.675660] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] self.wait() [ 551.675660] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 551.675660] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] self[:] = self._gt.wait() [ 551.675660] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 551.675660] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] return self._exit_event.wait() [ 551.675660] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 551.675660] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] result = hub.switch() [ 551.676102] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 551.676102] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] return self.greenlet.switch() [ 551.676102] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 551.676102] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] result = function(*args, **kwargs) [ 551.676102] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 551.676102] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] return func(*args, **kwargs) [ 551.676102] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 551.676102] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] raise e [ 551.676102] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 551.676102] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] nwinfo = self.network_api.allocate_for_instance( [ 551.676102] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 551.676102] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] created_port_ids = self._update_ports_for_instance( [ 551.676102] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 551.677359] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] with excutils.save_and_reraise_exception(): [ 551.677359] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 551.677359] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] self.force_reraise() [ 551.677359] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 551.677359] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] raise self.value [ 551.677359] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 551.677359] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] updated_port = self._update_port( [ 551.677359] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 551.677359] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] _ensure_no_port_binding_failure(port) [ 551.677359] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 551.677359] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] raise exception.PortBindingFailed(port_id=port['id']) [ 551.677359] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] nova.exception.PortBindingFailed: Binding failed for port 946407e1-ac24-4d57-9f2f-f2c0994ff966, please check neutron logs for more information. [ 551.677359] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] [ 551.677702] env[63088]: INFO nova.compute.manager [None req-96b6b7c1-255c-445f-b8db-9f4849d061fb tempest-ImagesNegativeTestJSON-1434405362 tempest-ImagesNegativeTestJSON-1434405362-project-member] [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] Terminating instance [ 551.682187] env[63088]: DEBUG oslo_concurrency.lockutils [None req-96b6b7c1-255c-445f-b8db-9f4849d061fb tempest-ImagesNegativeTestJSON-1434405362 tempest-ImagesNegativeTestJSON-1434405362-project-member] Acquiring lock "refresh_cache-0fd65b6b-2363-4b45-bc07-365157a8b052" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 551.682187] env[63088]: DEBUG oslo_concurrency.lockutils [None req-96b6b7c1-255c-445f-b8db-9f4849d061fb tempest-ImagesNegativeTestJSON-1434405362 tempest-ImagesNegativeTestJSON-1434405362-project-member] Acquired lock "refresh_cache-0fd65b6b-2363-4b45-bc07-365157a8b052" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 551.682404] env[63088]: DEBUG nova.network.neutron [None req-96b6b7c1-255c-445f-b8db-9f4849d061fb tempest-ImagesNegativeTestJSON-1434405362 tempest-ImagesNegativeTestJSON-1434405362-project-member] [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 552.129999] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 552.135502] env[63088]: DEBUG nova.compute.manager [None req-6e35ce22-cf37-483a-8093-9574de222716 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 552.144345] env[63088]: DEBUG nova.compute.utils [None req-710306d6-422f-4cdd-91a1-33da212e5f03 tempest-ServersAdminNegativeTestJSON-229078574 tempest-ServersAdminNegativeTestJSON-229078574-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 552.145630] env[63088]: DEBUG nova.compute.manager [None req-710306d6-422f-4cdd-91a1-33da212e5f03 tempest-ServersAdminNegativeTestJSON-229078574 tempest-ServersAdminNegativeTestJSON-229078574-project-member] [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 552.145810] env[63088]: DEBUG nova.network.neutron [None req-710306d6-422f-4cdd-91a1-33da212e5f03 tempest-ServersAdminNegativeTestJSON-229078574 tempest-ServersAdminNegativeTestJSON-229078574-project-member] [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 552.247280] env[63088]: DEBUG nova.network.neutron [None req-96b6b7c1-255c-445f-b8db-9f4849d061fb tempest-ImagesNegativeTestJSON-1434405362 tempest-ImagesNegativeTestJSON-1434405362-project-member] [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 552.305128] env[63088]: DEBUG nova.policy [None req-710306d6-422f-4cdd-91a1-33da212e5f03 tempest-ServersAdminNegativeTestJSON-229078574 tempest-ServersAdminNegativeTestJSON-229078574-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a4e1a7e5e6254f298593ec4dc67b9c2f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3e0736f71889430a813a05f2afccf6e2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 552.529857] env[63088]: DEBUG nova.network.neutron [None req-96b6b7c1-255c-445f-b8db-9f4849d061fb tempest-ImagesNegativeTestJSON-1434405362 tempest-ImagesNegativeTestJSON-1434405362-project-member] [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 552.616027] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 552.616686] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 552.616686] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Starting heal instance info cache {{(pid=63088) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 552.616686] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Rebuilding the list of instances to heal {{(pid=63088) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 552.654140] env[63088]: DEBUG nova.compute.manager [None req-710306d6-422f-4cdd-91a1-33da212e5f03 tempest-ServersAdminNegativeTestJSON-229078574 tempest-ServersAdminNegativeTestJSON-229078574-project-member] [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 552.670227] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6e35ce22-cf37-483a-8093-9574de222716 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 552.871394] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78097d84-595a-4d39-a1b1-1ac10fad17dc {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.881366] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-433f9f6e-8848-4c7d-926f-ee7420adf48c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.916387] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b04ea5a-9c13-4585-a56f-74cc519755cb {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.926805] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07d0bda4-5b4f-4f69-829f-1d3fde9fabfd {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.941718] env[63088]: DEBUG nova.compute.provider_tree [None req-8a34ea86-0280-447d-8082-40a2090a8af5 tempest-ServerDiagnosticsTest-1360445823 tempest-ServerDiagnosticsTest-1360445823-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 552.953293] env[63088]: DEBUG nova.compute.manager [req-54cc8956-80b3-4274-8f4e-36232a7b8270 req-85f21b8a-4b85-467c-9d9b-4b4620c415b4 service nova] [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] Received event network-changed-946407e1-ac24-4d57-9f2f-f2c0994ff966 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 552.953502] env[63088]: DEBUG nova.compute.manager [req-54cc8956-80b3-4274-8f4e-36232a7b8270 req-85f21b8a-4b85-467c-9d9b-4b4620c415b4 service nova] [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] Refreshing instance network info cache due to event network-changed-946407e1-ac24-4d57-9f2f-f2c0994ff966. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 552.953807] env[63088]: DEBUG oslo_concurrency.lockutils [req-54cc8956-80b3-4274-8f4e-36232a7b8270 req-85f21b8a-4b85-467c-9d9b-4b4620c415b4 service nova] Acquiring lock "refresh_cache-0fd65b6b-2363-4b45-bc07-365157a8b052" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 553.033189] env[63088]: DEBUG oslo_concurrency.lockutils [None req-96b6b7c1-255c-445f-b8db-9f4849d061fb tempest-ImagesNegativeTestJSON-1434405362 tempest-ImagesNegativeTestJSON-1434405362-project-member] Releasing lock "refresh_cache-0fd65b6b-2363-4b45-bc07-365157a8b052" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 553.034021] env[63088]: DEBUG nova.compute.manager [None req-96b6b7c1-255c-445f-b8db-9f4849d061fb tempest-ImagesNegativeTestJSON-1434405362 tempest-ImagesNegativeTestJSON-1434405362-project-member] [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 553.034021] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-96b6b7c1-255c-445f-b8db-9f4849d061fb tempest-ImagesNegativeTestJSON-1434405362 tempest-ImagesNegativeTestJSON-1434405362-project-member] [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 553.034130] env[63088]: DEBUG oslo_concurrency.lockutils [req-54cc8956-80b3-4274-8f4e-36232a7b8270 req-85f21b8a-4b85-467c-9d9b-4b4620c415b4 service nova] Acquired lock "refresh_cache-0fd65b6b-2363-4b45-bc07-365157a8b052" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 553.034549] env[63088]: DEBUG nova.network.neutron [req-54cc8956-80b3-4274-8f4e-36232a7b8270 req-85f21b8a-4b85-467c-9d9b-4b4620c415b4 service nova] [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] Refreshing network info cache for port 946407e1-ac24-4d57-9f2f-f2c0994ff966 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 553.035331] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e59e79fe-e686-42b8-bdc6-ec8634329de8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.054211] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65614918-7bbb-45b9-8a2b-eea15e0dd6ef {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.093193] env[63088]: WARNING nova.virt.vmwareapi.vmops [None req-96b6b7c1-255c-445f-b8db-9f4849d061fb tempest-ImagesNegativeTestJSON-1434405362 tempest-ImagesNegativeTestJSON-1434405362-project-member] [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0fd65b6b-2363-4b45-bc07-365157a8b052 could not be found. [ 553.094223] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-96b6b7c1-255c-445f-b8db-9f4849d061fb tempest-ImagesNegativeTestJSON-1434405362 tempest-ImagesNegativeTestJSON-1434405362-project-member] [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 553.094830] env[63088]: INFO nova.compute.manager [None req-96b6b7c1-255c-445f-b8db-9f4849d061fb tempest-ImagesNegativeTestJSON-1434405362 tempest-ImagesNegativeTestJSON-1434405362-project-member] [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] Took 0.06 seconds to destroy the instance on the hypervisor. [ 553.095362] env[63088]: DEBUG oslo.service.loopingcall [None req-96b6b7c1-255c-445f-b8db-9f4849d061fb tempest-ImagesNegativeTestJSON-1434405362 tempest-ImagesNegativeTestJSON-1434405362-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 553.095977] env[63088]: DEBUG nova.compute.manager [-] [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 553.098399] env[63088]: DEBUG nova.network.neutron [-] [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 553.123590] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] Skipping network cache update for instance because it is Building. {{(pid=63088) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 553.126038] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] Skipping network cache update for instance because it is Building. {{(pid=63088) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 553.126038] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] Skipping network cache update for instance because it is Building. {{(pid=63088) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 553.126038] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] Skipping network cache update for instance because it is Building. {{(pid=63088) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 553.126038] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Didn't find any instances for network info cache update. {{(pid=63088) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10012}} [ 553.126038] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 553.126038] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 553.126285] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 553.126285] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 553.126285] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 553.126285] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 553.126285] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63088) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 553.126285] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager.update_available_resource {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 553.131600] env[63088]: DEBUG nova.network.neutron [-] [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 553.289531] env[63088]: DEBUG nova.network.neutron [None req-710306d6-422f-4cdd-91a1-33da212e5f03 tempest-ServersAdminNegativeTestJSON-229078574 tempest-ServersAdminNegativeTestJSON-229078574-project-member] [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] Successfully created port: f91b4261-8234-46b0-900f-271a33cad644 {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 553.447395] env[63088]: DEBUG nova.scheduler.client.report [None req-8a34ea86-0280-447d-8082-40a2090a8af5 tempest-ServerDiagnosticsTest-1360445823 tempest-ServerDiagnosticsTest-1360445823-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 553.562426] env[63088]: DEBUG nova.network.neutron [req-54cc8956-80b3-4274-8f4e-36232a7b8270 req-85f21b8a-4b85-467c-9d9b-4b4620c415b4 service nova] [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 553.641057] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 553.641531] env[63088]: DEBUG nova.network.neutron [-] [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 553.656409] env[63088]: DEBUG nova.network.neutron [req-54cc8956-80b3-4274-8f4e-36232a7b8270 req-85f21b8a-4b85-467c-9d9b-4b4620c415b4 service nova] [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 553.675374] env[63088]: DEBUG nova.compute.manager [None req-710306d6-422f-4cdd-91a1-33da212e5f03 tempest-ServersAdminNegativeTestJSON-229078574 tempest-ServersAdminNegativeTestJSON-229078574-project-member] [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 553.705913] env[63088]: DEBUG nova.virt.hardware [None req-710306d6-422f-4cdd-91a1-33da212e5f03 tempest-ServersAdminNegativeTestJSON-229078574 tempest-ServersAdminNegativeTestJSON-229078574-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 553.705913] env[63088]: DEBUG nova.virt.hardware [None req-710306d6-422f-4cdd-91a1-33da212e5f03 tempest-ServersAdminNegativeTestJSON-229078574 tempest-ServersAdminNegativeTestJSON-229078574-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 553.706109] env[63088]: DEBUG nova.virt.hardware [None req-710306d6-422f-4cdd-91a1-33da212e5f03 tempest-ServersAdminNegativeTestJSON-229078574 tempest-ServersAdminNegativeTestJSON-229078574-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 553.706202] env[63088]: DEBUG nova.virt.hardware [None req-710306d6-422f-4cdd-91a1-33da212e5f03 tempest-ServersAdminNegativeTestJSON-229078574 tempest-ServersAdminNegativeTestJSON-229078574-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 553.706309] env[63088]: DEBUG nova.virt.hardware [None req-710306d6-422f-4cdd-91a1-33da212e5f03 tempest-ServersAdminNegativeTestJSON-229078574 tempest-ServersAdminNegativeTestJSON-229078574-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 553.706548] env[63088]: DEBUG nova.virt.hardware [None req-710306d6-422f-4cdd-91a1-33da212e5f03 tempest-ServersAdminNegativeTestJSON-229078574 tempest-ServersAdminNegativeTestJSON-229078574-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 553.706548] env[63088]: DEBUG nova.virt.hardware [None req-710306d6-422f-4cdd-91a1-33da212e5f03 tempest-ServersAdminNegativeTestJSON-229078574 tempest-ServersAdminNegativeTestJSON-229078574-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 553.707029] env[63088]: DEBUG nova.virt.hardware [None req-710306d6-422f-4cdd-91a1-33da212e5f03 tempest-ServersAdminNegativeTestJSON-229078574 tempest-ServersAdminNegativeTestJSON-229078574-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 553.707029] env[63088]: DEBUG nova.virt.hardware [None req-710306d6-422f-4cdd-91a1-33da212e5f03 tempest-ServersAdminNegativeTestJSON-229078574 tempest-ServersAdminNegativeTestJSON-229078574-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 553.707373] env[63088]: DEBUG nova.virt.hardware [None req-710306d6-422f-4cdd-91a1-33da212e5f03 tempest-ServersAdminNegativeTestJSON-229078574 tempest-ServersAdminNegativeTestJSON-229078574-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 553.707373] env[63088]: DEBUG nova.virt.hardware [None req-710306d6-422f-4cdd-91a1-33da212e5f03 tempest-ServersAdminNegativeTestJSON-229078574 tempest-ServersAdminNegativeTestJSON-229078574-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 553.709384] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20a26681-2b07-4b8a-b323-4ee6636ca88b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.716781] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1dbe7cc-0fc7-4a4a-96b6-e46069dd651b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.813983] env[63088]: ERROR nova.compute.manager [None req-fa54bfad-a3c9-4ca3-8f38-8bc8499d1488 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f66bce24-6dbd-4128-9ab8-86cf1c74d872, please check neutron logs for more information. [ 553.813983] env[63088]: ERROR nova.compute.manager Traceback (most recent call last): [ 553.813983] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 553.813983] env[63088]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 553.813983] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 553.813983] env[63088]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 553.813983] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 553.813983] env[63088]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 553.813983] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 553.813983] env[63088]: ERROR nova.compute.manager self.force_reraise() [ 553.813983] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 553.813983] env[63088]: ERROR nova.compute.manager raise self.value [ 553.813983] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 553.813983] env[63088]: ERROR nova.compute.manager updated_port = self._update_port( [ 553.813983] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 553.813983] env[63088]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 553.814590] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 553.814590] env[63088]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 553.814590] env[63088]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f66bce24-6dbd-4128-9ab8-86cf1c74d872, please check neutron logs for more information. [ 553.814590] env[63088]: ERROR nova.compute.manager [ 553.814590] env[63088]: Traceback (most recent call last): [ 553.814590] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 553.814590] env[63088]: listener.cb(fileno) [ 553.814590] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 553.814590] env[63088]: result = function(*args, **kwargs) [ 553.814590] env[63088]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 553.814590] env[63088]: return func(*args, **kwargs) [ 553.814590] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 553.814590] env[63088]: raise e [ 553.814590] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 553.814590] env[63088]: nwinfo = self.network_api.allocate_for_instance( [ 553.814590] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 553.814590] env[63088]: created_port_ids = self._update_ports_for_instance( [ 553.814590] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 553.814590] env[63088]: with excutils.save_and_reraise_exception(): [ 553.814590] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 553.814590] env[63088]: self.force_reraise() [ 553.814590] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 553.814590] env[63088]: raise self.value [ 553.814590] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 553.814590] env[63088]: updated_port = self._update_port( [ 553.814590] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 553.814590] env[63088]: _ensure_no_port_binding_failure(port) [ 553.814590] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 553.814590] env[63088]: raise exception.PortBindingFailed(port_id=port['id']) [ 553.815803] env[63088]: nova.exception.PortBindingFailed: Binding failed for port f66bce24-6dbd-4128-9ab8-86cf1c74d872, please check neutron logs for more information. [ 553.815803] env[63088]: Removing descriptor: 16 [ 553.815803] env[63088]: ERROR nova.compute.manager [None req-fa54bfad-a3c9-4ca3-8f38-8bc8499d1488 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f66bce24-6dbd-4128-9ab8-86cf1c74d872, please check neutron logs for more information. [ 553.815803] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] Traceback (most recent call last): [ 553.815803] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 553.815803] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] yield resources [ 553.815803] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 553.815803] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] self.driver.spawn(context, instance, image_meta, [ 553.815803] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 553.815803] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 553.815803] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 553.815803] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] vm_ref = self.build_virtual_machine(instance, [ 553.816873] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 553.816873] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] vif_infos = vmwarevif.get_vif_info(self._session, [ 553.816873] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 553.816873] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] for vif in network_info: [ 553.816873] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 553.816873] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] return self._sync_wrapper(fn, *args, **kwargs) [ 553.816873] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 553.816873] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] self.wait() [ 553.816873] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 553.816873] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] self[:] = self._gt.wait() [ 553.816873] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 553.816873] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] return self._exit_event.wait() [ 553.816873] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 553.817492] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] result = hub.switch() [ 553.817492] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 553.817492] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] return self.greenlet.switch() [ 553.817492] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 553.817492] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] result = function(*args, **kwargs) [ 553.817492] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 553.817492] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] return func(*args, **kwargs) [ 553.817492] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 553.817492] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] raise e [ 553.817492] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 553.817492] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] nwinfo = self.network_api.allocate_for_instance( [ 553.817492] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 553.817492] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] created_port_ids = self._update_ports_for_instance( [ 553.818128] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 553.818128] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] with excutils.save_and_reraise_exception(): [ 553.818128] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 553.818128] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] self.force_reraise() [ 553.818128] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 553.818128] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] raise self.value [ 553.818128] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 553.818128] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] updated_port = self._update_port( [ 553.818128] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 553.818128] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] _ensure_no_port_binding_failure(port) [ 553.818128] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 553.818128] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] raise exception.PortBindingFailed(port_id=port['id']) [ 553.818706] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] nova.exception.PortBindingFailed: Binding failed for port f66bce24-6dbd-4128-9ab8-86cf1c74d872, please check neutron logs for more information. [ 553.818706] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] [ 553.818706] env[63088]: INFO nova.compute.manager [None req-fa54bfad-a3c9-4ca3-8f38-8bc8499d1488 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] Terminating instance [ 553.819860] env[63088]: DEBUG oslo_concurrency.lockutils [None req-fa54bfad-a3c9-4ca3-8f38-8bc8499d1488 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] Acquiring lock "refresh_cache-05d2ad93-949e-42e3-85e9-eed9baaebe8c" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 553.820044] env[63088]: DEBUG oslo_concurrency.lockutils [None req-fa54bfad-a3c9-4ca3-8f38-8bc8499d1488 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] Acquired lock "refresh_cache-05d2ad93-949e-42e3-85e9-eed9baaebe8c" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 553.820216] env[63088]: DEBUG nova.network.neutron [None req-fa54bfad-a3c9-4ca3-8f38-8bc8499d1488 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 553.951131] env[63088]: DEBUG oslo_concurrency.lockutils [None req-8a34ea86-0280-447d-8082-40a2090a8af5 tempest-ServerDiagnosticsTest-1360445823 tempest-ServerDiagnosticsTest-1360445823-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.308s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 553.951931] env[63088]: DEBUG nova.compute.manager [None req-8a34ea86-0280-447d-8082-40a2090a8af5 tempest-ServerDiagnosticsTest-1360445823 tempest-ServerDiagnosticsTest-1360445823-project-member] [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 553.954649] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.129s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 553.955954] env[63088]: INFO nova.compute.claims [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 554.144789] env[63088]: INFO nova.compute.manager [-] [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] Took 1.05 seconds to deallocate network for instance. [ 554.147957] env[63088]: DEBUG nova.compute.claims [None req-96b6b7c1-255c-445f-b8db-9f4849d061fb tempest-ImagesNegativeTestJSON-1434405362 tempest-ImagesNegativeTestJSON-1434405362-project-member] [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] Aborting claim: {{(pid=63088) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 554.147957] env[63088]: DEBUG oslo_concurrency.lockutils [None req-96b6b7c1-255c-445f-b8db-9f4849d061fb tempest-ImagesNegativeTestJSON-1434405362 tempest-ImagesNegativeTestJSON-1434405362-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 554.165633] env[63088]: DEBUG oslo_concurrency.lockutils [req-54cc8956-80b3-4274-8f4e-36232a7b8270 req-85f21b8a-4b85-467c-9d9b-4b4620c415b4 service nova] Releasing lock "refresh_cache-0fd65b6b-2363-4b45-bc07-365157a8b052" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 554.356620] env[63088]: DEBUG nova.network.neutron [None req-fa54bfad-a3c9-4ca3-8f38-8bc8499d1488 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 554.460679] env[63088]: DEBUG nova.compute.utils [None req-8a34ea86-0280-447d-8082-40a2090a8af5 tempest-ServerDiagnosticsTest-1360445823 tempest-ServerDiagnosticsTest-1360445823-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 554.465811] env[63088]: DEBUG nova.compute.manager [None req-8a34ea86-0280-447d-8082-40a2090a8af5 tempest-ServerDiagnosticsTest-1360445823 tempest-ServerDiagnosticsTest-1360445823-project-member] [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 554.466072] env[63088]: DEBUG nova.network.neutron [None req-8a34ea86-0280-447d-8082-40a2090a8af5 tempest-ServerDiagnosticsTest-1360445823 tempest-ServerDiagnosticsTest-1360445823-project-member] [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 554.629376] env[63088]: DEBUG nova.network.neutron [None req-fa54bfad-a3c9-4ca3-8f38-8bc8499d1488 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 554.639147] env[63088]: DEBUG nova.policy [None req-8a34ea86-0280-447d-8082-40a2090a8af5 tempest-ServerDiagnosticsTest-1360445823 tempest-ServerDiagnosticsTest-1360445823-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '79aefd034b8940b7bd27466ef0b672bd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4f25764741bc4c90a73f342b7f2a530b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 554.914177] env[63088]: ERROR nova.compute.manager [None req-710306d6-422f-4cdd-91a1-33da212e5f03 tempest-ServersAdminNegativeTestJSON-229078574 tempest-ServersAdminNegativeTestJSON-229078574-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f91b4261-8234-46b0-900f-271a33cad644, please check neutron logs for more information. [ 554.914177] env[63088]: ERROR nova.compute.manager Traceback (most recent call last): [ 554.914177] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 554.914177] env[63088]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 554.914177] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 554.914177] env[63088]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 554.914177] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 554.914177] env[63088]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 554.914177] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 554.914177] env[63088]: ERROR nova.compute.manager self.force_reraise() [ 554.914177] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 554.914177] env[63088]: ERROR nova.compute.manager raise self.value [ 554.914177] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 554.914177] env[63088]: ERROR nova.compute.manager updated_port = self._update_port( [ 554.914177] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 554.914177] env[63088]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 554.914569] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 554.914569] env[63088]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 554.914569] env[63088]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f91b4261-8234-46b0-900f-271a33cad644, please check neutron logs for more information. [ 554.914569] env[63088]: ERROR nova.compute.manager [ 554.914569] env[63088]: Traceback (most recent call last): [ 554.914569] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 554.914569] env[63088]: listener.cb(fileno) [ 554.914569] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 554.914569] env[63088]: result = function(*args, **kwargs) [ 554.914569] env[63088]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 554.914569] env[63088]: return func(*args, **kwargs) [ 554.914569] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 554.914569] env[63088]: raise e [ 554.914569] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 554.914569] env[63088]: nwinfo = self.network_api.allocate_for_instance( [ 554.914569] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 554.914569] env[63088]: created_port_ids = self._update_ports_for_instance( [ 554.914569] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 554.914569] env[63088]: with excutils.save_and_reraise_exception(): [ 554.914569] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 554.914569] env[63088]: self.force_reraise() [ 554.914569] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 554.914569] env[63088]: raise self.value [ 554.914569] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 554.914569] env[63088]: updated_port = self._update_port( [ 554.914569] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 554.914569] env[63088]: _ensure_no_port_binding_failure(port) [ 554.914569] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 554.914569] env[63088]: raise exception.PortBindingFailed(port_id=port['id']) [ 554.915262] env[63088]: nova.exception.PortBindingFailed: Binding failed for port f91b4261-8234-46b0-900f-271a33cad644, please check neutron logs for more information. [ 554.915262] env[63088]: Removing descriptor: 15 [ 554.915428] env[63088]: ERROR nova.compute.manager [None req-710306d6-422f-4cdd-91a1-33da212e5f03 tempest-ServersAdminNegativeTestJSON-229078574 tempest-ServersAdminNegativeTestJSON-229078574-project-member] [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f91b4261-8234-46b0-900f-271a33cad644, please check neutron logs for more information. [ 554.915428] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] Traceback (most recent call last): [ 554.915428] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 554.915428] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] yield resources [ 554.915428] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 554.915428] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] self.driver.spawn(context, instance, image_meta, [ 554.915428] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 554.915428] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] self._vmops.spawn(context, instance, image_meta, injected_files, [ 554.915428] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 554.915428] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] vm_ref = self.build_virtual_machine(instance, [ 554.915428] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 554.915763] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] vif_infos = vmwarevif.get_vif_info(self._session, [ 554.915763] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 554.915763] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] for vif in network_info: [ 554.915763] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 554.915763] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] return self._sync_wrapper(fn, *args, **kwargs) [ 554.915763] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 554.915763] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] self.wait() [ 554.915763] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 554.915763] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] self[:] = self._gt.wait() [ 554.915763] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 554.915763] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] return self._exit_event.wait() [ 554.915763] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 554.915763] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] result = hub.switch() [ 554.916151] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 554.916151] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] return self.greenlet.switch() [ 554.916151] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 554.916151] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] result = function(*args, **kwargs) [ 554.916151] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 554.916151] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] return func(*args, **kwargs) [ 554.916151] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 554.916151] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] raise e [ 554.916151] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 554.916151] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] nwinfo = self.network_api.allocate_for_instance( [ 554.916151] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 554.916151] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] created_port_ids = self._update_ports_for_instance( [ 554.916151] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 554.916606] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] with excutils.save_and_reraise_exception(): [ 554.916606] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 554.916606] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] self.force_reraise() [ 554.916606] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 554.916606] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] raise self.value [ 554.916606] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 554.916606] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] updated_port = self._update_port( [ 554.916606] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 554.916606] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] _ensure_no_port_binding_failure(port) [ 554.916606] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 554.916606] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] raise exception.PortBindingFailed(port_id=port['id']) [ 554.916606] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] nova.exception.PortBindingFailed: Binding failed for port f91b4261-8234-46b0-900f-271a33cad644, please check neutron logs for more information. [ 554.916606] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] [ 554.917884] env[63088]: INFO nova.compute.manager [None req-710306d6-422f-4cdd-91a1-33da212e5f03 tempest-ServersAdminNegativeTestJSON-229078574 tempest-ServersAdminNegativeTestJSON-229078574-project-member] [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] Terminating instance [ 554.918545] env[63088]: DEBUG oslo_concurrency.lockutils [None req-710306d6-422f-4cdd-91a1-33da212e5f03 tempest-ServersAdminNegativeTestJSON-229078574 tempest-ServersAdminNegativeTestJSON-229078574-project-member] Acquiring lock "refresh_cache-4f487d47-daf7-4ef1-9e8b-8d8f870bfe52" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 554.918742] env[63088]: DEBUG oslo_concurrency.lockutils [None req-710306d6-422f-4cdd-91a1-33da212e5f03 tempest-ServersAdminNegativeTestJSON-229078574 tempest-ServersAdminNegativeTestJSON-229078574-project-member] Acquired lock "refresh_cache-4f487d47-daf7-4ef1-9e8b-8d8f870bfe52" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 554.918917] env[63088]: DEBUG nova.network.neutron [None req-710306d6-422f-4cdd-91a1-33da212e5f03 tempest-ServersAdminNegativeTestJSON-229078574 tempest-ServersAdminNegativeTestJSON-229078574-project-member] [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 554.970220] env[63088]: DEBUG nova.compute.manager [None req-8a34ea86-0280-447d-8082-40a2090a8af5 tempest-ServerDiagnosticsTest-1360445823 tempest-ServerDiagnosticsTest-1360445823-project-member] [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 555.132568] env[63088]: DEBUG oslo_concurrency.lockutils [None req-fa54bfad-a3c9-4ca3-8f38-8bc8499d1488 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] Releasing lock "refresh_cache-05d2ad93-949e-42e3-85e9-eed9baaebe8c" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 555.133103] env[63088]: DEBUG nova.compute.manager [None req-fa54bfad-a3c9-4ca3-8f38-8bc8499d1488 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 555.133372] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-fa54bfad-a3c9-4ca3-8f38-8bc8499d1488 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 555.134764] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-80560cad-b552-4ec2-b0bf-76c5771b858a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.137490] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26bc06a1-3bc3-4ed6-bc69-50ce2f07852e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.149243] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-932ca1ce-e047-4b01-a98c-b5de49880c9f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.168328] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-985d7a3f-5cd3-4a04-8de8-314c5667f839 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.180676] env[63088]: WARNING nova.virt.vmwareapi.vmops [None req-fa54bfad-a3c9-4ca3-8f38-8bc8499d1488 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 05d2ad93-949e-42e3-85e9-eed9baaebe8c could not be found. [ 555.180676] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-fa54bfad-a3c9-4ca3-8f38-8bc8499d1488 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 555.180676] env[63088]: INFO nova.compute.manager [None req-fa54bfad-a3c9-4ca3-8f38-8bc8499d1488 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] Took 0.05 seconds to destroy the instance on the hypervisor. [ 555.180676] env[63088]: DEBUG oslo.service.loopingcall [None req-fa54bfad-a3c9-4ca3-8f38-8bc8499d1488 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 555.180676] env[63088]: DEBUG nova.compute.manager [-] [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 555.181163] env[63088]: DEBUG nova.network.neutron [-] [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 555.210786] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59345c05-e0fc-455d-b88f-b4a8071eaeb6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.218958] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad5f6204-3e23-4be5-b8ba-f1582c7fda35 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.234173] env[63088]: DEBUG nova.compute.provider_tree [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 555.239725] env[63088]: DEBUG nova.network.neutron [-] [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 555.458796] env[63088]: DEBUG nova.network.neutron [None req-710306d6-422f-4cdd-91a1-33da212e5f03 tempest-ServersAdminNegativeTestJSON-229078574 tempest-ServersAdminNegativeTestJSON-229078574-project-member] [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 555.641160] env[63088]: DEBUG nova.network.neutron [None req-710306d6-422f-4cdd-91a1-33da212e5f03 tempest-ServersAdminNegativeTestJSON-229078574 tempest-ServersAdminNegativeTestJSON-229078574-project-member] [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 555.742374] env[63088]: DEBUG nova.scheduler.client.report [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 555.748806] env[63088]: DEBUG nova.network.neutron [-] [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 555.802998] env[63088]: DEBUG nova.network.neutron [None req-8a34ea86-0280-447d-8082-40a2090a8af5 tempest-ServerDiagnosticsTest-1360445823 tempest-ServerDiagnosticsTest-1360445823-project-member] [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] Successfully created port: 784e2da7-96ad-48a1-9570-f494c3f4b49a {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 555.982129] env[63088]: DEBUG nova.compute.manager [None req-8a34ea86-0280-447d-8082-40a2090a8af5 tempest-ServerDiagnosticsTest-1360445823 tempest-ServerDiagnosticsTest-1360445823-project-member] [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 556.019635] env[63088]: DEBUG nova.virt.hardware [None req-8a34ea86-0280-447d-8082-40a2090a8af5 tempest-ServerDiagnosticsTest-1360445823 tempest-ServerDiagnosticsTest-1360445823-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 556.020869] env[63088]: DEBUG nova.virt.hardware [None req-8a34ea86-0280-447d-8082-40a2090a8af5 tempest-ServerDiagnosticsTest-1360445823 tempest-ServerDiagnosticsTest-1360445823-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 556.020869] env[63088]: DEBUG nova.virt.hardware [None req-8a34ea86-0280-447d-8082-40a2090a8af5 tempest-ServerDiagnosticsTest-1360445823 tempest-ServerDiagnosticsTest-1360445823-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 556.020869] env[63088]: DEBUG nova.virt.hardware [None req-8a34ea86-0280-447d-8082-40a2090a8af5 tempest-ServerDiagnosticsTest-1360445823 tempest-ServerDiagnosticsTest-1360445823-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 556.020869] env[63088]: DEBUG nova.virt.hardware [None req-8a34ea86-0280-447d-8082-40a2090a8af5 tempest-ServerDiagnosticsTest-1360445823 tempest-ServerDiagnosticsTest-1360445823-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 556.020869] env[63088]: DEBUG nova.virt.hardware [None req-8a34ea86-0280-447d-8082-40a2090a8af5 tempest-ServerDiagnosticsTest-1360445823 tempest-ServerDiagnosticsTest-1360445823-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 556.021196] env[63088]: DEBUG nova.virt.hardware [None req-8a34ea86-0280-447d-8082-40a2090a8af5 tempest-ServerDiagnosticsTest-1360445823 tempest-ServerDiagnosticsTest-1360445823-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 556.021196] env[63088]: DEBUG nova.virt.hardware [None req-8a34ea86-0280-447d-8082-40a2090a8af5 tempest-ServerDiagnosticsTest-1360445823 tempest-ServerDiagnosticsTest-1360445823-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 556.021196] env[63088]: DEBUG nova.virt.hardware [None req-8a34ea86-0280-447d-8082-40a2090a8af5 tempest-ServerDiagnosticsTest-1360445823 tempest-ServerDiagnosticsTest-1360445823-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 556.023196] env[63088]: DEBUG nova.virt.hardware [None req-8a34ea86-0280-447d-8082-40a2090a8af5 tempest-ServerDiagnosticsTest-1360445823 tempest-ServerDiagnosticsTest-1360445823-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 556.023196] env[63088]: DEBUG nova.virt.hardware [None req-8a34ea86-0280-447d-8082-40a2090a8af5 tempest-ServerDiagnosticsTest-1360445823 tempest-ServerDiagnosticsTest-1360445823-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 556.023196] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21c0534f-6d40-4b43-8eef-9ab2978bd4b4 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.034457] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57d746b7-3cba-4ac5-a7a3-f33dc1a412e3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.144182] env[63088]: DEBUG oslo_concurrency.lockutils [None req-710306d6-422f-4cdd-91a1-33da212e5f03 tempest-ServersAdminNegativeTestJSON-229078574 tempest-ServersAdminNegativeTestJSON-229078574-project-member] Releasing lock "refresh_cache-4f487d47-daf7-4ef1-9e8b-8d8f870bfe52" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 556.144557] env[63088]: DEBUG nova.compute.manager [None req-710306d6-422f-4cdd-91a1-33da212e5f03 tempest-ServersAdminNegativeTestJSON-229078574 tempest-ServersAdminNegativeTestJSON-229078574-project-member] [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 556.147719] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-710306d6-422f-4cdd-91a1-33da212e5f03 tempest-ServersAdminNegativeTestJSON-229078574 tempest-ServersAdminNegativeTestJSON-229078574-project-member] [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 556.147719] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-67067707-cfb7-4417-8c52-d189cc994b1d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.158059] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0fe26dc-e189-4119-a16b-9e70a2f49f7f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.186985] env[63088]: WARNING nova.virt.vmwareapi.vmops [None req-710306d6-422f-4cdd-91a1-33da212e5f03 tempest-ServersAdminNegativeTestJSON-229078574 tempest-ServersAdminNegativeTestJSON-229078574-project-member] [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52 could not be found. [ 556.187103] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-710306d6-422f-4cdd-91a1-33da212e5f03 tempest-ServersAdminNegativeTestJSON-229078574 tempest-ServersAdminNegativeTestJSON-229078574-project-member] [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 556.187289] env[63088]: INFO nova.compute.manager [None req-710306d6-422f-4cdd-91a1-33da212e5f03 tempest-ServersAdminNegativeTestJSON-229078574 tempest-ServersAdminNegativeTestJSON-229078574-project-member] [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] Took 0.04 seconds to destroy the instance on the hypervisor. [ 556.187560] env[63088]: DEBUG oslo.service.loopingcall [None req-710306d6-422f-4cdd-91a1-33da212e5f03 tempest-ServersAdminNegativeTestJSON-229078574 tempest-ServersAdminNegativeTestJSON-229078574-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 556.187794] env[63088]: DEBUG nova.compute.manager [-] [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 556.187936] env[63088]: DEBUG nova.network.neutron [-] [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 556.213911] env[63088]: DEBUG nova.network.neutron [-] [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 556.251750] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.297s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 556.252591] env[63088]: DEBUG nova.compute.manager [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 556.256602] env[63088]: INFO nova.compute.manager [-] [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] Took 1.08 seconds to deallocate network for instance. [ 556.256894] env[63088]: DEBUG oslo_concurrency.lockutils [None req-61dd383c-8cef-4a29-a464-69dfd03cb5ce tempest-ServerExternalEventsTest-515168927 tempest-ServerExternalEventsTest-515168927-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.659s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 556.258462] env[63088]: INFO nova.compute.claims [None req-61dd383c-8cef-4a29-a464-69dfd03cb5ce tempest-ServerExternalEventsTest-515168927 tempest-ServerExternalEventsTest-515168927-project-member] [instance: 173fccf1-f258-4f0c-a968-466a27406c55] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 556.266811] env[63088]: DEBUG nova.compute.claims [None req-fa54bfad-a3c9-4ca3-8f38-8bc8499d1488 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] Aborting claim: {{(pid=63088) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 556.266811] env[63088]: DEBUG oslo_concurrency.lockutils [None req-fa54bfad-a3c9-4ca3-8f38-8bc8499d1488 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 556.717369] env[63088]: DEBUG nova.network.neutron [-] [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 556.767233] env[63088]: DEBUG nova.compute.utils [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 556.779317] env[63088]: DEBUG nova.compute.manager [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Not allocating networking since 'none' was specified. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 556.787347] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d1272b04-ee5c-4e9e-b0bf-6ca9d0629aae tempest-VolumesAssistedSnapshotsTest-1004656420 tempest-VolumesAssistedSnapshotsTest-1004656420-project-member] Acquiring lock "d2c07175-b635-4900-9dab-9767e5f6bce4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 556.787347] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d1272b04-ee5c-4e9e-b0bf-6ca9d0629aae tempest-VolumesAssistedSnapshotsTest-1004656420 tempest-VolumesAssistedSnapshotsTest-1004656420-project-member] Lock "d2c07175-b635-4900-9dab-9767e5f6bce4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 557.092159] env[63088]: DEBUG nova.compute.manager [req-1533209f-0b7d-40e9-9f42-0e347489f652 req-980b3fd1-9c61-45d8-8f5d-0056336054dc service nova] [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] Received event network-vif-deleted-946407e1-ac24-4d57-9f2f-f2c0994ff966 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 557.092159] env[63088]: DEBUG nova.compute.manager [req-1533209f-0b7d-40e9-9f42-0e347489f652 req-980b3fd1-9c61-45d8-8f5d-0056336054dc service nova] [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] Received event network-changed-f66bce24-6dbd-4128-9ab8-86cf1c74d872 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 557.092159] env[63088]: DEBUG nova.compute.manager [req-1533209f-0b7d-40e9-9f42-0e347489f652 req-980b3fd1-9c61-45d8-8f5d-0056336054dc service nova] [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] Refreshing instance network info cache due to event network-changed-f66bce24-6dbd-4128-9ab8-86cf1c74d872. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 557.092159] env[63088]: DEBUG oslo_concurrency.lockutils [req-1533209f-0b7d-40e9-9f42-0e347489f652 req-980b3fd1-9c61-45d8-8f5d-0056336054dc service nova] Acquiring lock "refresh_cache-05d2ad93-949e-42e3-85e9-eed9baaebe8c" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 557.092159] env[63088]: DEBUG oslo_concurrency.lockutils [req-1533209f-0b7d-40e9-9f42-0e347489f652 req-980b3fd1-9c61-45d8-8f5d-0056336054dc service nova] Acquired lock "refresh_cache-05d2ad93-949e-42e3-85e9-eed9baaebe8c" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 557.092314] env[63088]: DEBUG nova.network.neutron [req-1533209f-0b7d-40e9-9f42-0e347489f652 req-980b3fd1-9c61-45d8-8f5d-0056336054dc service nova] [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] Refreshing network info cache for port f66bce24-6dbd-4128-9ab8-86cf1c74d872 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 557.222656] env[63088]: INFO nova.compute.manager [-] [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] Took 1.03 seconds to deallocate network for instance. [ 557.224937] env[63088]: DEBUG nova.compute.claims [None req-710306d6-422f-4cdd-91a1-33da212e5f03 tempest-ServersAdminNegativeTestJSON-229078574 tempest-ServersAdminNegativeTestJSON-229078574-project-member] [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] Aborting claim: {{(pid=63088) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 557.225130] env[63088]: DEBUG oslo_concurrency.lockutils [None req-710306d6-422f-4cdd-91a1-33da212e5f03 tempest-ServersAdminNegativeTestJSON-229078574 tempest-ServersAdminNegativeTestJSON-229078574-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 557.280739] env[63088]: DEBUG nova.compute.manager [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 557.459018] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91addaa2-69c0-4213-a34c-5bfe59fc3efe {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.466335] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4efa269e-4477-4808-a2a0-96a0388a0cd8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.498285] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b15b3ea-9746-4d0f-9ae3-0cc5fefcc54a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.507138] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d7bbae8-7822-4886-b314-32a819dc7887 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.522174] env[63088]: DEBUG nova.compute.provider_tree [None req-61dd383c-8cef-4a29-a464-69dfd03cb5ce tempest-ServerExternalEventsTest-515168927 tempest-ServerExternalEventsTest-515168927-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 557.640522] env[63088]: DEBUG nova.network.neutron [req-1533209f-0b7d-40e9-9f42-0e347489f652 req-980b3fd1-9c61-45d8-8f5d-0056336054dc service nova] [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 557.795098] env[63088]: DEBUG nova.network.neutron [req-1533209f-0b7d-40e9-9f42-0e347489f652 req-980b3fd1-9c61-45d8-8f5d-0056336054dc service nova] [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 558.025480] env[63088]: DEBUG nova.scheduler.client.report [None req-61dd383c-8cef-4a29-a464-69dfd03cb5ce tempest-ServerExternalEventsTest-515168927 tempest-ServerExternalEventsTest-515168927-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 558.299107] env[63088]: DEBUG nova.compute.manager [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 558.305711] env[63088]: DEBUG oslo_concurrency.lockutils [req-1533209f-0b7d-40e9-9f42-0e347489f652 req-980b3fd1-9c61-45d8-8f5d-0056336054dc service nova] Releasing lock "refresh_cache-05d2ad93-949e-42e3-85e9-eed9baaebe8c" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 558.305952] env[63088]: DEBUG nova.compute.manager [req-1533209f-0b7d-40e9-9f42-0e347489f652 req-980b3fd1-9c61-45d8-8f5d-0056336054dc service nova] [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] Received event network-vif-deleted-f66bce24-6dbd-4128-9ab8-86cf1c74d872 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 558.306143] env[63088]: DEBUG nova.compute.manager [req-1533209f-0b7d-40e9-9f42-0e347489f652 req-980b3fd1-9c61-45d8-8f5d-0056336054dc service nova] [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] Received event network-changed-f91b4261-8234-46b0-900f-271a33cad644 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 558.306302] env[63088]: DEBUG nova.compute.manager [req-1533209f-0b7d-40e9-9f42-0e347489f652 req-980b3fd1-9c61-45d8-8f5d-0056336054dc service nova] [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] Refreshing instance network info cache due to event network-changed-f91b4261-8234-46b0-900f-271a33cad644. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 558.306496] env[63088]: DEBUG oslo_concurrency.lockutils [req-1533209f-0b7d-40e9-9f42-0e347489f652 req-980b3fd1-9c61-45d8-8f5d-0056336054dc service nova] Acquiring lock "refresh_cache-4f487d47-daf7-4ef1-9e8b-8d8f870bfe52" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 558.306633] env[63088]: DEBUG oslo_concurrency.lockutils [req-1533209f-0b7d-40e9-9f42-0e347489f652 req-980b3fd1-9c61-45d8-8f5d-0056336054dc service nova] Acquired lock "refresh_cache-4f487d47-daf7-4ef1-9e8b-8d8f870bfe52" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 558.306792] env[63088]: DEBUG nova.network.neutron [req-1533209f-0b7d-40e9-9f42-0e347489f652 req-980b3fd1-9c61-45d8-8f5d-0056336054dc service nova] [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] Refreshing network info cache for port f91b4261-8234-46b0-900f-271a33cad644 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 558.334415] env[63088]: DEBUG nova.virt.hardware [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 558.334657] env[63088]: DEBUG nova.virt.hardware [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 558.334812] env[63088]: DEBUG nova.virt.hardware [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 558.334984] env[63088]: DEBUG nova.virt.hardware [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 558.335131] env[63088]: DEBUG nova.virt.hardware [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 558.336955] env[63088]: DEBUG nova.virt.hardware [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 558.336955] env[63088]: DEBUG nova.virt.hardware [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 558.336955] env[63088]: DEBUG nova.virt.hardware [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 558.336955] env[63088]: DEBUG nova.virt.hardware [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 558.336955] env[63088]: DEBUG nova.virt.hardware [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 558.338120] env[63088]: DEBUG nova.virt.hardware [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 558.340019] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b545b7a-9137-4cc0-b18d-130170542050 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.354069] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a394031d-3687-4f38-9be3-660a9165ddd6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.369242] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Instance VIF info [] {{(pid=63088) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 558.378772] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Creating folder: OpenStack. Parent ref: group-v4. {{(pid=63088) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 558.379036] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1c55717e-0698-4804-b257-80733b3ccc8a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.392730] env[63088]: INFO nova.virt.vmwareapi.vm_util [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Created folder: OpenStack in parent group-v4. [ 558.392922] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Creating folder: Project (ace9922ecaac4a6fb811bd5f51dba841). Parent ref: group-v275816. {{(pid=63088) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 558.393243] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5f805696-5edb-4955-80f3-b07e17b67bf1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.405223] env[63088]: INFO nova.virt.vmwareapi.vm_util [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Created folder: Project (ace9922ecaac4a6fb811bd5f51dba841) in parent group-v275816. [ 558.405223] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Creating folder: Instances. Parent ref: group-v275817. {{(pid=63088) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 558.405223] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2f7f37c6-1e32-43d5-95f3-8eac954a76d5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.415022] env[63088]: INFO nova.virt.vmwareapi.vm_util [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Created folder: Instances in parent group-v275817. [ 558.415022] env[63088]: DEBUG oslo.service.loopingcall [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 558.415022] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Creating VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 558.415022] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d623fcde-5ec9-4f0f-919a-963e1a1c97cf {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.438282] env[63088]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 558.438282] env[63088]: value = "task-1284763" [ 558.438282] env[63088]: _type = "Task" [ 558.438282] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 558.444739] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1284763, 'name': CreateVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 558.533460] env[63088]: DEBUG oslo_concurrency.lockutils [None req-61dd383c-8cef-4a29-a464-69dfd03cb5ce tempest-ServerExternalEventsTest-515168927 tempest-ServerExternalEventsTest-515168927-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.276s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 558.533977] env[63088]: DEBUG nova.compute.manager [None req-61dd383c-8cef-4a29-a464-69dfd03cb5ce tempest-ServerExternalEventsTest-515168927 tempest-ServerExternalEventsTest-515168927-project-member] [instance: 173fccf1-f258-4f0c-a968-466a27406c55] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 558.536551] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bc610745-9520-42a6-9fd1-7e4c9c63b77a tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.852s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 558.537964] env[63088]: INFO nova.compute.claims [None req-bc610745-9520-42a6-9fd1-7e4c9c63b77a tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 558.895039] env[63088]: DEBUG nova.network.neutron [req-1533209f-0b7d-40e9-9f42-0e347489f652 req-980b3fd1-9c61-45d8-8f5d-0056336054dc service nova] [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 558.951764] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1284763, 'name': CreateVM_Task, 'duration_secs': 0.347151} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 558.951937] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Created VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 558.954409] env[63088]: DEBUG oslo_vmware.service [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b1fa2a1-c963-498c-a3ac-fa2caa71232c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.962891] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 558.963063] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 558.963728] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 558.963981] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5023a382-9204-4a4e-8896-fc80f3ea34cc {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.971592] env[63088]: DEBUG oslo_vmware.api [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Waiting for the task: (returnval){ [ 558.971592] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]5271b2da-8ade-c1a0-1905-5f3f5dd8d460" [ 558.971592] env[63088]: _type = "Task" [ 558.971592] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 558.979782] env[63088]: DEBUG oslo_vmware.api [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]5271b2da-8ade-c1a0-1905-5f3f5dd8d460, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 559.044176] env[63088]: DEBUG nova.compute.utils [None req-61dd383c-8cef-4a29-a464-69dfd03cb5ce tempest-ServerExternalEventsTest-515168927 tempest-ServerExternalEventsTest-515168927-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 559.053018] env[63088]: DEBUG nova.compute.manager [None req-61dd383c-8cef-4a29-a464-69dfd03cb5ce tempest-ServerExternalEventsTest-515168927 tempest-ServerExternalEventsTest-515168927-project-member] [instance: 173fccf1-f258-4f0c-a968-466a27406c55] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 559.053018] env[63088]: DEBUG nova.network.neutron [None req-61dd383c-8cef-4a29-a464-69dfd03cb5ce tempest-ServerExternalEventsTest-515168927 tempest-ServerExternalEventsTest-515168927-project-member] [instance: 173fccf1-f258-4f0c-a968-466a27406c55] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 559.159651] env[63088]: DEBUG nova.network.neutron [req-1533209f-0b7d-40e9-9f42-0e347489f652 req-980b3fd1-9c61-45d8-8f5d-0056336054dc service nova] [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 559.303354] env[63088]: DEBUG nova.policy [None req-61dd383c-8cef-4a29-a464-69dfd03cb5ce tempest-ServerExternalEventsTest-515168927 tempest-ServerExternalEventsTest-515168927-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3ef85b6ec4d4430aae63e5ad5da0c8ba', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3b58739093a64b3f91370ad5de5d38e5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 559.488328] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 559.488841] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Processing image 1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 559.489205] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 559.489462] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 559.490082] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 559.490542] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8efa6b8a-2f9c-4d12-b737-23788ceea745 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.509694] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 559.509923] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 559.513469] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6986f5f-78eb-46ff-95a7-e1056b032a98 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.526134] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-844d5147-27ed-4e19-b534-113f708d69e4 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.531159] env[63088]: DEBUG oslo_vmware.api [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Waiting for the task: (returnval){ [ 559.531159] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52035b4c-47fd-8fc5-72aa-c00d13ac5b51" [ 559.531159] env[63088]: _type = "Task" [ 559.531159] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 559.542371] env[63088]: DEBUG oslo_vmware.api [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52035b4c-47fd-8fc5-72aa-c00d13ac5b51, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 559.555785] env[63088]: DEBUG nova.compute.manager [None req-61dd383c-8cef-4a29-a464-69dfd03cb5ce tempest-ServerExternalEventsTest-515168927 tempest-ServerExternalEventsTest-515168927-project-member] [instance: 173fccf1-f258-4f0c-a968-466a27406c55] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 559.664686] env[63088]: DEBUG oslo_concurrency.lockutils [req-1533209f-0b7d-40e9-9f42-0e347489f652 req-980b3fd1-9c61-45d8-8f5d-0056336054dc service nova] Releasing lock "refresh_cache-4f487d47-daf7-4ef1-9e8b-8d8f870bfe52" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 559.665017] env[63088]: DEBUG nova.compute.manager [req-1533209f-0b7d-40e9-9f42-0e347489f652 req-980b3fd1-9c61-45d8-8f5d-0056336054dc service nova] [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] Received event network-vif-deleted-f91b4261-8234-46b0-900f-271a33cad644 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 559.871859] env[63088]: ERROR nova.compute.manager [None req-8a34ea86-0280-447d-8082-40a2090a8af5 tempest-ServerDiagnosticsTest-1360445823 tempest-ServerDiagnosticsTest-1360445823-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 784e2da7-96ad-48a1-9570-f494c3f4b49a, please check neutron logs for more information. [ 559.871859] env[63088]: ERROR nova.compute.manager Traceback (most recent call last): [ 559.871859] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 559.871859] env[63088]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 559.871859] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 559.871859] env[63088]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 559.871859] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 559.871859] env[63088]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 559.871859] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 559.871859] env[63088]: ERROR nova.compute.manager self.force_reraise() [ 559.871859] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 559.871859] env[63088]: ERROR nova.compute.manager raise self.value [ 559.871859] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 559.871859] env[63088]: ERROR nova.compute.manager updated_port = self._update_port( [ 559.871859] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 559.871859] env[63088]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 559.872591] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 559.872591] env[63088]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 559.872591] env[63088]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 784e2da7-96ad-48a1-9570-f494c3f4b49a, please check neutron logs for more information. [ 559.872591] env[63088]: ERROR nova.compute.manager [ 559.872591] env[63088]: Traceback (most recent call last): [ 559.872591] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 559.872591] env[63088]: listener.cb(fileno) [ 559.872591] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 559.872591] env[63088]: result = function(*args, **kwargs) [ 559.872591] env[63088]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 559.872591] env[63088]: return func(*args, **kwargs) [ 559.872591] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 559.872591] env[63088]: raise e [ 559.872591] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 559.872591] env[63088]: nwinfo = self.network_api.allocate_for_instance( [ 559.872591] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 559.872591] env[63088]: created_port_ids = self._update_ports_for_instance( [ 559.872591] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 559.872591] env[63088]: with excutils.save_and_reraise_exception(): [ 559.872591] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 559.872591] env[63088]: self.force_reraise() [ 559.872591] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 559.872591] env[63088]: raise self.value [ 559.872591] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 559.872591] env[63088]: updated_port = self._update_port( [ 559.872591] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 559.872591] env[63088]: _ensure_no_port_binding_failure(port) [ 559.872591] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 559.872591] env[63088]: raise exception.PortBindingFailed(port_id=port['id']) [ 559.874870] env[63088]: nova.exception.PortBindingFailed: Binding failed for port 784e2da7-96ad-48a1-9570-f494c3f4b49a, please check neutron logs for more information. [ 559.874870] env[63088]: Removing descriptor: 17 [ 559.874870] env[63088]: ERROR nova.compute.manager [None req-8a34ea86-0280-447d-8082-40a2090a8af5 tempest-ServerDiagnosticsTest-1360445823 tempest-ServerDiagnosticsTest-1360445823-project-member] [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 784e2da7-96ad-48a1-9570-f494c3f4b49a, please check neutron logs for more information. [ 559.874870] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] Traceback (most recent call last): [ 559.874870] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 559.874870] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] yield resources [ 559.874870] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 559.874870] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] self.driver.spawn(context, instance, image_meta, [ 559.874870] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 559.874870] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] self._vmops.spawn(context, instance, image_meta, injected_files, [ 559.874870] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 559.874870] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] vm_ref = self.build_virtual_machine(instance, [ 559.875622] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 559.875622] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] vif_infos = vmwarevif.get_vif_info(self._session, [ 559.875622] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 559.875622] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] for vif in network_info: [ 559.875622] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 559.875622] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] return self._sync_wrapper(fn, *args, **kwargs) [ 559.875622] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 559.875622] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] self.wait() [ 559.875622] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 559.875622] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] self[:] = self._gt.wait() [ 559.875622] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 559.875622] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] return self._exit_event.wait() [ 559.875622] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 559.875964] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] result = hub.switch() [ 559.875964] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 559.875964] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] return self.greenlet.switch() [ 559.875964] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 559.875964] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] result = function(*args, **kwargs) [ 559.875964] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 559.875964] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] return func(*args, **kwargs) [ 559.875964] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 559.875964] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] raise e [ 559.875964] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 559.875964] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] nwinfo = self.network_api.allocate_for_instance( [ 559.875964] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 559.875964] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] created_port_ids = self._update_ports_for_instance( [ 559.876329] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 559.876329] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] with excutils.save_and_reraise_exception(): [ 559.876329] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 559.876329] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] self.force_reraise() [ 559.876329] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 559.876329] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] raise self.value [ 559.876329] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 559.876329] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] updated_port = self._update_port( [ 559.876329] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 559.876329] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] _ensure_no_port_binding_failure(port) [ 559.876329] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 559.876329] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] raise exception.PortBindingFailed(port_id=port['id']) [ 559.876646] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] nova.exception.PortBindingFailed: Binding failed for port 784e2da7-96ad-48a1-9570-f494c3f4b49a, please check neutron logs for more information. [ 559.876646] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] [ 559.876646] env[63088]: INFO nova.compute.manager [None req-8a34ea86-0280-447d-8082-40a2090a8af5 tempest-ServerDiagnosticsTest-1360445823 tempest-ServerDiagnosticsTest-1360445823-project-member] [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] Terminating instance [ 559.878878] env[63088]: DEBUG oslo_concurrency.lockutils [None req-8a34ea86-0280-447d-8082-40a2090a8af5 tempest-ServerDiagnosticsTest-1360445823 tempest-ServerDiagnosticsTest-1360445823-project-member] Acquiring lock "refresh_cache-2d7bf024-afb5-4d33-9c5e-06633167ac77" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 559.879133] env[63088]: DEBUG oslo_concurrency.lockutils [None req-8a34ea86-0280-447d-8082-40a2090a8af5 tempest-ServerDiagnosticsTest-1360445823 tempest-ServerDiagnosticsTest-1360445823-project-member] Acquired lock "refresh_cache-2d7bf024-afb5-4d33-9c5e-06633167ac77" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 559.879373] env[63088]: DEBUG nova.network.neutron [None req-8a34ea86-0280-447d-8082-40a2090a8af5 tempest-ServerDiagnosticsTest-1360445823 tempest-ServerDiagnosticsTest-1360445823-project-member] [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 559.913708] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-649b3b92-b8ff-4e6f-8f75-d11d8be44c4b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.922868] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06a3b7ca-ef52-4677-b716-d98e8bed0a2f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.955481] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-518768a9-712d-4e5a-8e76-2b26d7831f2e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.963586] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-494308ad-e267-42af-a553-68989ab93bfe {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.981720] env[63088]: DEBUG nova.compute.provider_tree [None req-bc610745-9520-42a6-9fd1-7e4c9c63b77a tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 560.043371] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Preparing fetch location {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 560.043696] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Creating directory with path [datastore1] vmware_temp/a4e62e3b-e26b-4d44-908f-d1f2ad366dbb/1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 560.043938] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c6b5b3aa-7822-4984-9cc9-28850c83b230 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.071322] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Created directory with path [datastore1] vmware_temp/a4e62e3b-e26b-4d44-908f-d1f2ad366dbb/1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 560.073016] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Fetch image to [datastore1] vmware_temp/a4e62e3b-e26b-4d44-908f-d1f2ad366dbb/1e8c5d18-0a03-4e18-afe1-de5a6e255953/tmp-sparse.vmdk {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 560.073016] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Downloading image file data 1e8c5d18-0a03-4e18-afe1-de5a6e255953 to [datastore1] vmware_temp/a4e62e3b-e26b-4d44-908f-d1f2ad366dbb/1e8c5d18-0a03-4e18-afe1-de5a6e255953/tmp-sparse.vmdk on the data store datastore1 {{(pid=63088) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 560.073016] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c1def5d-af3c-468a-a3ce-9d88afa58c3f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.080373] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58939bb5-0e01-4fb0-b7f2-f380a745ad6d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.093826] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71683941-0bfd-40c0-964c-1deb5dfd2bf6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.128902] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f193fdf4-d4ab-499e-b823-ea5c9a44ea36 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.136469] env[63088]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-6deeaa4d-b827-4564-bba5-70d87f0f73d5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.229282] env[63088]: DEBUG nova.virt.vmwareapi.images [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Downloading image file data 1e8c5d18-0a03-4e18-afe1-de5a6e255953 to the data store datastore1 {{(pid=63088) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 560.340113] env[63088]: DEBUG oslo_vmware.rw_handles [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/a4e62e3b-e26b-4d44-908f-d1f2ad366dbb/1e8c5d18-0a03-4e18-afe1-de5a6e255953/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=63088) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 560.477333] env[63088]: DEBUG nova.network.neutron [None req-8a34ea86-0280-447d-8082-40a2090a8af5 tempest-ServerDiagnosticsTest-1360445823 tempest-ServerDiagnosticsTest-1360445823-project-member] [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 560.485187] env[63088]: DEBUG nova.scheduler.client.report [None req-bc610745-9520-42a6-9fd1-7e4c9c63b77a tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 560.572361] env[63088]: DEBUG nova.compute.manager [None req-61dd383c-8cef-4a29-a464-69dfd03cb5ce tempest-ServerExternalEventsTest-515168927 tempest-ServerExternalEventsTest-515168927-project-member] [instance: 173fccf1-f258-4f0c-a968-466a27406c55] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 560.600697] env[63088]: DEBUG nova.virt.hardware [None req-61dd383c-8cef-4a29-a464-69dfd03cb5ce tempest-ServerExternalEventsTest-515168927 tempest-ServerExternalEventsTest-515168927-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 560.600697] env[63088]: DEBUG nova.virt.hardware [None req-61dd383c-8cef-4a29-a464-69dfd03cb5ce tempest-ServerExternalEventsTest-515168927 tempest-ServerExternalEventsTest-515168927-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 560.600697] env[63088]: DEBUG nova.virt.hardware [None req-61dd383c-8cef-4a29-a464-69dfd03cb5ce tempest-ServerExternalEventsTest-515168927 tempest-ServerExternalEventsTest-515168927-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 560.600969] env[63088]: DEBUG nova.virt.hardware [None req-61dd383c-8cef-4a29-a464-69dfd03cb5ce tempest-ServerExternalEventsTest-515168927 tempest-ServerExternalEventsTest-515168927-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 560.600969] env[63088]: DEBUG nova.virt.hardware [None req-61dd383c-8cef-4a29-a464-69dfd03cb5ce tempest-ServerExternalEventsTest-515168927 tempest-ServerExternalEventsTest-515168927-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 560.600969] env[63088]: DEBUG nova.virt.hardware [None req-61dd383c-8cef-4a29-a464-69dfd03cb5ce tempest-ServerExternalEventsTest-515168927 tempest-ServerExternalEventsTest-515168927-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 560.601076] env[63088]: DEBUG nova.virt.hardware [None req-61dd383c-8cef-4a29-a464-69dfd03cb5ce tempest-ServerExternalEventsTest-515168927 tempest-ServerExternalEventsTest-515168927-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 560.601161] env[63088]: DEBUG nova.virt.hardware [None req-61dd383c-8cef-4a29-a464-69dfd03cb5ce tempest-ServerExternalEventsTest-515168927 tempest-ServerExternalEventsTest-515168927-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 560.601362] env[63088]: DEBUG nova.virt.hardware [None req-61dd383c-8cef-4a29-a464-69dfd03cb5ce tempest-ServerExternalEventsTest-515168927 tempest-ServerExternalEventsTest-515168927-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 560.601486] env[63088]: DEBUG nova.virt.hardware [None req-61dd383c-8cef-4a29-a464-69dfd03cb5ce tempest-ServerExternalEventsTest-515168927 tempest-ServerExternalEventsTest-515168927-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 560.601730] env[63088]: DEBUG nova.virt.hardware [None req-61dd383c-8cef-4a29-a464-69dfd03cb5ce tempest-ServerExternalEventsTest-515168927 tempest-ServerExternalEventsTest-515168927-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 560.602734] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dbb3358-6218-42d2-8a8a-2d65195905e1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.614567] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd76cc2a-143e-4c32-9396-80f8313bc9b2 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.621731] env[63088]: DEBUG nova.compute.manager [req-949869a3-be99-404b-9c7f-3db2895cf09b req-a474839a-2c1d-4a47-80ba-a332c3aa2aac service nova] [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] Received event network-changed-784e2da7-96ad-48a1-9570-f494c3f4b49a {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 560.621998] env[63088]: DEBUG nova.compute.manager [req-949869a3-be99-404b-9c7f-3db2895cf09b req-a474839a-2c1d-4a47-80ba-a332c3aa2aac service nova] [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] Refreshing instance network info cache due to event network-changed-784e2da7-96ad-48a1-9570-f494c3f4b49a. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 560.622224] env[63088]: DEBUG oslo_concurrency.lockutils [req-949869a3-be99-404b-9c7f-3db2895cf09b req-a474839a-2c1d-4a47-80ba-a332c3aa2aac service nova] Acquiring lock "refresh_cache-2d7bf024-afb5-4d33-9c5e-06633167ac77" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 560.693927] env[63088]: DEBUG nova.network.neutron [None req-8a34ea86-0280-447d-8082-40a2090a8af5 tempest-ServerDiagnosticsTest-1360445823 tempest-ServerDiagnosticsTest-1360445823-project-member] [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 560.992362] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bc610745-9520-42a6-9fd1-7e4c9c63b77a tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.454s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 560.992362] env[63088]: DEBUG nova.compute.manager [None req-bc610745-9520-42a6-9fd1-7e4c9c63b77a tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 560.994069] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6568dcbe-ebf7-46f4-8cbf-a6cec6f2c264 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.777s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 560.999393] env[63088]: INFO nova.compute.claims [None req-6568dcbe-ebf7-46f4-8cbf-a6cec6f2c264 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] [instance: 4d5293a0-6945-4731-afb6-996a142c8447] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 561.120940] env[63088]: DEBUG nova.network.neutron [None req-61dd383c-8cef-4a29-a464-69dfd03cb5ce tempest-ServerExternalEventsTest-515168927 tempest-ServerExternalEventsTest-515168927-project-member] [instance: 173fccf1-f258-4f0c-a968-466a27406c55] Successfully created port: 06dbc948-3bcb-4bb1-9083-7a5a312da444 {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 561.197072] env[63088]: DEBUG oslo_concurrency.lockutils [None req-8a34ea86-0280-447d-8082-40a2090a8af5 tempest-ServerDiagnosticsTest-1360445823 tempest-ServerDiagnosticsTest-1360445823-project-member] Releasing lock "refresh_cache-2d7bf024-afb5-4d33-9c5e-06633167ac77" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 561.198143] env[63088]: DEBUG nova.compute.manager [None req-8a34ea86-0280-447d-8082-40a2090a8af5 tempest-ServerDiagnosticsTest-1360445823 tempest-ServerDiagnosticsTest-1360445823-project-member] [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 561.198143] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-8a34ea86-0280-447d-8082-40a2090a8af5 tempest-ServerDiagnosticsTest-1360445823 tempest-ServerDiagnosticsTest-1360445823-project-member] [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 561.198143] env[63088]: DEBUG oslo_concurrency.lockutils [req-949869a3-be99-404b-9c7f-3db2895cf09b req-a474839a-2c1d-4a47-80ba-a332c3aa2aac service nova] Acquired lock "refresh_cache-2d7bf024-afb5-4d33-9c5e-06633167ac77" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 561.198242] env[63088]: DEBUG nova.network.neutron [req-949869a3-be99-404b-9c7f-3db2895cf09b req-a474839a-2c1d-4a47-80ba-a332c3aa2aac service nova] [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] Refreshing network info cache for port 784e2da7-96ad-48a1-9570-f494c3f4b49a {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 561.201030] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8056a919-9e6c-4ce0-8f96-083865b8b14d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.213510] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8973ba9b-9fd6-4608-84ba-56968b5026cd {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.242722] env[63088]: WARNING nova.virt.vmwareapi.vmops [None req-8a34ea86-0280-447d-8082-40a2090a8af5 tempest-ServerDiagnosticsTest-1360445823 tempest-ServerDiagnosticsTest-1360445823-project-member] [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2d7bf024-afb5-4d33-9c5e-06633167ac77 could not be found. [ 561.242939] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-8a34ea86-0280-447d-8082-40a2090a8af5 tempest-ServerDiagnosticsTest-1360445823 tempest-ServerDiagnosticsTest-1360445823-project-member] [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 561.243216] env[63088]: INFO nova.compute.manager [None req-8a34ea86-0280-447d-8082-40a2090a8af5 tempest-ServerDiagnosticsTest-1360445823 tempest-ServerDiagnosticsTest-1360445823-project-member] [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] Took 0.05 seconds to destroy the instance on the hypervisor. [ 561.243487] env[63088]: DEBUG oslo.service.loopingcall [None req-8a34ea86-0280-447d-8082-40a2090a8af5 tempest-ServerDiagnosticsTest-1360445823 tempest-ServerDiagnosticsTest-1360445823-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 561.245900] env[63088]: DEBUG nova.compute.manager [-] [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 561.246107] env[63088]: DEBUG nova.network.neutron [-] [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 561.292392] env[63088]: DEBUG nova.network.neutron [-] [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 561.459438] env[63088]: DEBUG oslo_vmware.rw_handles [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Completed reading data from the image iterator. {{(pid=63088) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 561.459438] env[63088]: DEBUG oslo_vmware.rw_handles [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Closing write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/a4e62e3b-e26b-4d44-908f-d1f2ad366dbb/1e8c5d18-0a03-4e18-afe1-de5a6e255953/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=63088) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 561.503720] env[63088]: DEBUG nova.compute.utils [None req-bc610745-9520-42a6-9fd1-7e4c9c63b77a tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 561.508202] env[63088]: DEBUG nova.compute.manager [None req-bc610745-9520-42a6-9fd1-7e4c9c63b77a tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 561.508393] env[63088]: DEBUG nova.network.neutron [None req-bc610745-9520-42a6-9fd1-7e4c9c63b77a tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 561.578905] env[63088]: DEBUG nova.policy [None req-bc610745-9520-42a6-9fd1-7e4c9c63b77a tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b9356a34eb6446339640348c78fa5f24', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2805c642e0d245cf92424e6a807c230e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 561.645223] env[63088]: DEBUG nova.virt.vmwareapi.images [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Downloaded image file data 1e8c5d18-0a03-4e18-afe1-de5a6e255953 to vmware_temp/a4e62e3b-e26b-4d44-908f-d1f2ad366dbb/1e8c5d18-0a03-4e18-afe1-de5a6e255953/tmp-sparse.vmdk on the data store datastore1 {{(pid=63088) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 561.645223] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Caching image {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 561.646263] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Copying Virtual Disk [datastore1] vmware_temp/a4e62e3b-e26b-4d44-908f-d1f2ad366dbb/1e8c5d18-0a03-4e18-afe1-de5a6e255953/tmp-sparse.vmdk to [datastore1] vmware_temp/a4e62e3b-e26b-4d44-908f-d1f2ad366dbb/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 561.646263] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-351d6b6b-773a-4ec1-af85-c0c9483cc4d8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.655580] env[63088]: DEBUG oslo_vmware.api [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Waiting for the task: (returnval){ [ 561.655580] env[63088]: value = "task-1284764" [ 561.655580] env[63088]: _type = "Task" [ 561.655580] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 561.665466] env[63088]: DEBUG oslo_vmware.api [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Task: {'id': task-1284764, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 561.766219] env[63088]: DEBUG nova.network.neutron [req-949869a3-be99-404b-9c7f-3db2895cf09b req-a474839a-2c1d-4a47-80ba-a332c3aa2aac service nova] [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 561.796174] env[63088]: DEBUG nova.network.neutron [-] [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 562.011921] env[63088]: DEBUG nova.compute.manager [None req-bc610745-9520-42a6-9fd1-7e4c9c63b77a tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 562.115402] env[63088]: DEBUG nova.network.neutron [req-949869a3-be99-404b-9c7f-3db2895cf09b req-a474839a-2c1d-4a47-80ba-a332c3aa2aac service nova] [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 562.171919] env[63088]: DEBUG oslo_vmware.api [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Task: {'id': task-1284764, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 562.225468] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-581b2313-7523-4493-abbf-b43d56f66def {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.234766] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc59a379-fb83-4fe7-ab42-1c0a51f60c7e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.268399] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c398341a-3909-4c70-b3df-aa473e6f4b01 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.280271] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94335583-796a-4c10-808a-59482f236601 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.295435] env[63088]: DEBUG nova.compute.provider_tree [None req-6568dcbe-ebf7-46f4-8cbf-a6cec6f2c264 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 562.301558] env[63088]: INFO nova.compute.manager [-] [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] Took 1.06 seconds to deallocate network for instance. [ 562.306099] env[63088]: DEBUG nova.compute.claims [None req-8a34ea86-0280-447d-8082-40a2090a8af5 tempest-ServerDiagnosticsTest-1360445823 tempest-ServerDiagnosticsTest-1360445823-project-member] [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] Aborting claim: {{(pid=63088) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 562.306099] env[63088]: DEBUG oslo_concurrency.lockutils [None req-8a34ea86-0280-447d-8082-40a2090a8af5 tempest-ServerDiagnosticsTest-1360445823 tempest-ServerDiagnosticsTest-1360445823-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 562.524921] env[63088]: DEBUG nova.network.neutron [None req-bc610745-9520-42a6-9fd1-7e4c9c63b77a tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] Successfully created port: 010a5358-1268-49c2-b2b4-9525552d6fdc {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 562.619848] env[63088]: DEBUG oslo_concurrency.lockutils [req-949869a3-be99-404b-9c7f-3db2895cf09b req-a474839a-2c1d-4a47-80ba-a332c3aa2aac service nova] Releasing lock "refresh_cache-2d7bf024-afb5-4d33-9c5e-06633167ac77" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 562.671871] env[63088]: DEBUG oslo_vmware.api [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Task: {'id': task-1284764, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.782463} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 562.675012] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Copied Virtual Disk [datastore1] vmware_temp/a4e62e3b-e26b-4d44-908f-d1f2ad366dbb/1e8c5d18-0a03-4e18-afe1-de5a6e255953/tmp-sparse.vmdk to [datastore1] vmware_temp/a4e62e3b-e26b-4d44-908f-d1f2ad366dbb/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 562.675263] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Deleting the datastore file [datastore1] vmware_temp/a4e62e3b-e26b-4d44-908f-d1f2ad366dbb/1e8c5d18-0a03-4e18-afe1-de5a6e255953/tmp-sparse.vmdk {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 562.675535] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-12477dfd-ab8a-4aa4-a2d2-9ceb8ebd2a75 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.682673] env[63088]: DEBUG oslo_vmware.api [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Waiting for the task: (returnval){ [ 562.682673] env[63088]: value = "task-1284765" [ 562.682673] env[63088]: _type = "Task" [ 562.682673] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 562.694403] env[63088]: DEBUG oslo_vmware.api [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Task: {'id': task-1284765, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 562.799050] env[63088]: DEBUG nova.scheduler.client.report [None req-6568dcbe-ebf7-46f4-8cbf-a6cec6f2c264 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 562.893194] env[63088]: DEBUG oslo_concurrency.lockutils [None req-79832ead-48a0-4aad-a47e-8e38c2a687da tempest-ServersTestJSON-106593364 tempest-ServersTestJSON-106593364-project-member] Acquiring lock "34c01c2c-519b-44a1-aba2-c26e2c0b4737" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 562.893517] env[63088]: DEBUG oslo_concurrency.lockutils [None req-79832ead-48a0-4aad-a47e-8e38c2a687da tempest-ServersTestJSON-106593364 tempest-ServersTestJSON-106593364-project-member] Lock "34c01c2c-519b-44a1-aba2-c26e2c0b4737" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 563.020763] env[63088]: DEBUG nova.compute.manager [None req-bc610745-9520-42a6-9fd1-7e4c9c63b77a tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 563.060464] env[63088]: DEBUG nova.virt.hardware [None req-bc610745-9520-42a6-9fd1-7e4c9c63b77a tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 563.060736] env[63088]: DEBUG nova.virt.hardware [None req-bc610745-9520-42a6-9fd1-7e4c9c63b77a tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 563.060826] env[63088]: DEBUG nova.virt.hardware [None req-bc610745-9520-42a6-9fd1-7e4c9c63b77a tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 563.061097] env[63088]: DEBUG nova.virt.hardware [None req-bc610745-9520-42a6-9fd1-7e4c9c63b77a tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 563.061143] env[63088]: DEBUG nova.virt.hardware [None req-bc610745-9520-42a6-9fd1-7e4c9c63b77a tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 563.061396] env[63088]: DEBUG nova.virt.hardware [None req-bc610745-9520-42a6-9fd1-7e4c9c63b77a tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 563.061509] env[63088]: DEBUG nova.virt.hardware [None req-bc610745-9520-42a6-9fd1-7e4c9c63b77a tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 563.061664] env[63088]: DEBUG nova.virt.hardware [None req-bc610745-9520-42a6-9fd1-7e4c9c63b77a tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 563.061819] env[63088]: DEBUG nova.virt.hardware [None req-bc610745-9520-42a6-9fd1-7e4c9c63b77a tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 563.061967] env[63088]: DEBUG nova.virt.hardware [None req-bc610745-9520-42a6-9fd1-7e4c9c63b77a tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 563.062148] env[63088]: DEBUG nova.virt.hardware [None req-bc610745-9520-42a6-9fd1-7e4c9c63b77a tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 563.063122] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-001af0bc-7d68-40ca-880b-53d050e69b05 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.071754] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5988f240-fb3b-4af7-95ea-b7cfe3c61788 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.198461] env[63088]: DEBUG oslo_vmware.api [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Task: {'id': task-1284765, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.026266} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 563.198744] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 563.198956] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Moving file from [datastore1] vmware_temp/a4e62e3b-e26b-4d44-908f-d1f2ad366dbb/1e8c5d18-0a03-4e18-afe1-de5a6e255953 to [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953. {{(pid=63088) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 563.199184] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-22869858-e8e4-4005-993a-3237a22a0904 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.207970] env[63088]: DEBUG oslo_vmware.api [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Waiting for the task: (returnval){ [ 563.207970] env[63088]: value = "task-1284766" [ 563.207970] env[63088]: _type = "Task" [ 563.207970] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 563.221218] env[63088]: DEBUG oslo_vmware.api [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Task: {'id': task-1284766, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 563.305863] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6568dcbe-ebf7-46f4-8cbf-a6cec6f2c264 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.312s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 563.306401] env[63088]: DEBUG nova.compute.manager [None req-6568dcbe-ebf7-46f4-8cbf-a6cec6f2c264 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] [instance: 4d5293a0-6945-4731-afb6-996a142c8447] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 563.312262] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.181s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 563.312262] env[63088]: INFO nova.compute.claims [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] [instance: c4777721-3f65-455f-9973-c1ed0732de34] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 563.727918] env[63088]: DEBUG oslo_vmware.api [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Task: {'id': task-1284766, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.030917} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 563.728589] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] File moved {{(pid=63088) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 563.728589] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Cleaning up location [datastore1] vmware_temp/a4e62e3b-e26b-4d44-908f-d1f2ad366dbb {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 563.728750] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Deleting the datastore file [datastore1] vmware_temp/a4e62e3b-e26b-4d44-908f-d1f2ad366dbb {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 563.728979] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6828d3df-dffa-4188-b3eb-3cc18a912479 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.737160] env[63088]: DEBUG oslo_vmware.api [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Waiting for the task: (returnval){ [ 563.737160] env[63088]: value = "task-1284767" [ 563.737160] env[63088]: _type = "Task" [ 563.737160] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 563.741585] env[63088]: DEBUG nova.network.neutron [None req-bc610745-9520-42a6-9fd1-7e4c9c63b77a tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] Successfully created port: 568fb202-ee89-4cbf-a7c4-3d5533b7659a {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 563.751635] env[63088]: DEBUG oslo_vmware.api [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Task: {'id': task-1284767, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 563.819528] env[63088]: DEBUG nova.compute.utils [None req-6568dcbe-ebf7-46f4-8cbf-a6cec6f2c264 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 563.824156] env[63088]: DEBUG nova.compute.manager [None req-6568dcbe-ebf7-46f4-8cbf-a6cec6f2c264 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] [instance: 4d5293a0-6945-4731-afb6-996a142c8447] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 563.825334] env[63088]: DEBUG nova.network.neutron [None req-6568dcbe-ebf7-46f4-8cbf-a6cec6f2c264 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] [instance: 4d5293a0-6945-4731-afb6-996a142c8447] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 564.015535] env[63088]: DEBUG nova.policy [None req-6568dcbe-ebf7-46f4-8cbf-a6cec6f2c264 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '95f3328a40e34d89ae4b07be1c364cfc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd101ee89dcc143e1876d2217b09cef11', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 564.157441] env[63088]: DEBUG nova.compute.manager [req-781e477d-811f-4562-9f1e-b15b69a0a873 req-a4bf5f53-a4ec-4e9f-a44b-2955e299f99b service nova] [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] Received event network-vif-deleted-784e2da7-96ad-48a1-9570-f494c3f4b49a {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 564.249652] env[63088]: DEBUG oslo_vmware.api [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Task: {'id': task-1284767, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.029797} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 564.249652] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 564.250345] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a418be23-46ac-4101-9fff-581f7b1a07d6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.262737] env[63088]: DEBUG oslo_vmware.api [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Waiting for the task: (returnval){ [ 564.262737] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]526ba96f-555c-dbb8-280f-85c29d90bd2a" [ 564.262737] env[63088]: _type = "Task" [ 564.262737] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 564.277074] env[63088]: DEBUG oslo_vmware.api [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]526ba96f-555c-dbb8-280f-85c29d90bd2a, 'name': SearchDatastore_Task, 'duration_secs': 0.010658} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 564.277340] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 564.277595] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] e8cfbce4-cfb2-463e-9f38-78a058dab6e1/e8cfbce4-cfb2-463e-9f38-78a058dab6e1.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 564.278517] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d08c0e6b-9674-43f0-83e1-c2d111a7db8d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.289246] env[63088]: DEBUG oslo_vmware.api [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Waiting for the task: (returnval){ [ 564.289246] env[63088]: value = "task-1284768" [ 564.289246] env[63088]: _type = "Task" [ 564.289246] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 564.302087] env[63088]: DEBUG oslo_vmware.api [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Task: {'id': task-1284768, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 564.333101] env[63088]: DEBUG nova.compute.manager [None req-6568dcbe-ebf7-46f4-8cbf-a6cec6f2c264 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] [instance: 4d5293a0-6945-4731-afb6-996a142c8447] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 564.462447] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a9773ab2-f8b9-4102-8e41-5ee881a8637e tempest-FloatingIPsAssociationNegativeTestJSON-189613961 tempest-FloatingIPsAssociationNegativeTestJSON-189613961-project-member] Acquiring lock "9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 564.462685] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a9773ab2-f8b9-4102-8e41-5ee881a8637e tempest-FloatingIPsAssociationNegativeTestJSON-189613961 tempest-FloatingIPsAssociationNegativeTestJSON-189613961-project-member] Lock "9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 564.597648] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d23ded29-e6be-4357-b102-90ad1775eabf {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.606236] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6489ebe2-ad4d-441c-ba8c-cc9d71fdf5fa {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.645301] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac7d19be-f162-4657-bf68-2563f5eae0fa {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.653218] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cfd90f6-4719-4a8f-9886-33259c723af5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.666995] env[63088]: DEBUG nova.compute.provider_tree [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Updating inventory in ProviderTree for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 564.803244] env[63088]: DEBUG oslo_vmware.api [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Task: {'id': task-1284768, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 565.136916] env[63088]: DEBUG nova.network.neutron [None req-6568dcbe-ebf7-46f4-8cbf-a6cec6f2c264 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] [instance: 4d5293a0-6945-4731-afb6-996a142c8447] Successfully created port: 0c9febcc-5137-47a2-a0b9-df090f78c2e7 {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 565.192901] env[63088]: ERROR nova.scheduler.client.report [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] [req-30e80213-1967-43f8-8b4f-d1a7ebb74d29] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 6eae54a9-8831-40eb-bf54-4bc60d346b02. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-30e80213-1967-43f8-8b4f-d1a7ebb74d29"}]} [ 565.211721] env[63088]: DEBUG nova.scheduler.client.report [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Refreshing inventories for resource provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 565.228337] env[63088]: DEBUG nova.scheduler.client.report [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Updating ProviderTree inventory for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 565.228605] env[63088]: DEBUG nova.compute.provider_tree [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Updating inventory in ProviderTree for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 565.246171] env[63088]: DEBUG nova.scheduler.client.report [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Refreshing aggregate associations for resource provider 6eae54a9-8831-40eb-bf54-4bc60d346b02, aggregates: None {{(pid=63088) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 565.265455] env[63088]: DEBUG nova.scheduler.client.report [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Refreshing trait associations for resource provider 6eae54a9-8831-40eb-bf54-4bc60d346b02, traits: COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO {{(pid=63088) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 565.268366] env[63088]: DEBUG nova.network.neutron [None req-bc610745-9520-42a6-9fd1-7e4c9c63b77a tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] Successfully created port: 9f4495ae-acc6-4a65-9628-3d753fcd69ed {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 565.299523] env[63088]: DEBUG oslo_vmware.api [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Task: {'id': task-1284768, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.540419} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 565.300081] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] e8cfbce4-cfb2-463e-9f38-78a058dab6e1/e8cfbce4-cfb2-463e-9f38-78a058dab6e1.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 565.300417] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Extending root virtual disk to 1048576 {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 565.300669] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1c27bef9-e361-4535-80bd-c1a69086087a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.309924] env[63088]: DEBUG oslo_vmware.api [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Waiting for the task: (returnval){ [ 565.309924] env[63088]: value = "task-1284769" [ 565.309924] env[63088]: _type = "Task" [ 565.309924] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 565.320950] env[63088]: DEBUG oslo_vmware.api [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Task: {'id': task-1284769, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 565.349584] env[63088]: DEBUG nova.compute.manager [None req-6568dcbe-ebf7-46f4-8cbf-a6cec6f2c264 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] [instance: 4d5293a0-6945-4731-afb6-996a142c8447] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 565.387314] env[63088]: DEBUG nova.virt.hardware [None req-6568dcbe-ebf7-46f4-8cbf-a6cec6f2c264 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 565.387643] env[63088]: DEBUG nova.virt.hardware [None req-6568dcbe-ebf7-46f4-8cbf-a6cec6f2c264 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 565.387834] env[63088]: DEBUG nova.virt.hardware [None req-6568dcbe-ebf7-46f4-8cbf-a6cec6f2c264 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 565.390516] env[63088]: DEBUG nova.virt.hardware [None req-6568dcbe-ebf7-46f4-8cbf-a6cec6f2c264 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 565.390516] env[63088]: DEBUG nova.virt.hardware [None req-6568dcbe-ebf7-46f4-8cbf-a6cec6f2c264 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 565.390516] env[63088]: DEBUG nova.virt.hardware [None req-6568dcbe-ebf7-46f4-8cbf-a6cec6f2c264 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 565.390516] env[63088]: DEBUG nova.virt.hardware [None req-6568dcbe-ebf7-46f4-8cbf-a6cec6f2c264 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 565.390516] env[63088]: DEBUG nova.virt.hardware [None req-6568dcbe-ebf7-46f4-8cbf-a6cec6f2c264 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 565.390781] env[63088]: DEBUG nova.virt.hardware [None req-6568dcbe-ebf7-46f4-8cbf-a6cec6f2c264 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 565.390781] env[63088]: DEBUG nova.virt.hardware [None req-6568dcbe-ebf7-46f4-8cbf-a6cec6f2c264 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 565.390781] env[63088]: DEBUG nova.virt.hardware [None req-6568dcbe-ebf7-46f4-8cbf-a6cec6f2c264 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 565.391726] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-032f1364-4cd1-4ff6-975d-2e57670146fc {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.404271] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5969fafb-b0d7-4caf-ac4d-1abba3e9e6aa {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.497576] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-351e1080-5546-405c-8819-54c9b7ec654f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.505429] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75e73d13-f33b-46f7-8b47-731b48af6298 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.538041] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4ed4934-2eee-4713-ae55-75eaa5153d8b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.543448] env[63088]: ERROR nova.compute.manager [None req-61dd383c-8cef-4a29-a464-69dfd03cb5ce tempest-ServerExternalEventsTest-515168927 tempest-ServerExternalEventsTest-515168927-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 06dbc948-3bcb-4bb1-9083-7a5a312da444, please check neutron logs for more information. [ 565.543448] env[63088]: ERROR nova.compute.manager Traceback (most recent call last): [ 565.543448] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 565.543448] env[63088]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 565.543448] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 565.543448] env[63088]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 565.543448] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 565.543448] env[63088]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 565.543448] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 565.543448] env[63088]: ERROR nova.compute.manager self.force_reraise() [ 565.543448] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 565.543448] env[63088]: ERROR nova.compute.manager raise self.value [ 565.543448] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 565.543448] env[63088]: ERROR nova.compute.manager updated_port = self._update_port( [ 565.543448] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 565.543448] env[63088]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 565.544150] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 565.544150] env[63088]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 565.544150] env[63088]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 06dbc948-3bcb-4bb1-9083-7a5a312da444, please check neutron logs for more information. [ 565.544150] env[63088]: ERROR nova.compute.manager [ 565.544150] env[63088]: Traceback (most recent call last): [ 565.544150] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 565.544150] env[63088]: listener.cb(fileno) [ 565.544150] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 565.544150] env[63088]: result = function(*args, **kwargs) [ 565.544150] env[63088]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 565.544150] env[63088]: return func(*args, **kwargs) [ 565.544150] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 565.544150] env[63088]: raise e [ 565.544150] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 565.544150] env[63088]: nwinfo = self.network_api.allocate_for_instance( [ 565.544150] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 565.544150] env[63088]: created_port_ids = self._update_ports_for_instance( [ 565.544150] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 565.544150] env[63088]: with excutils.save_and_reraise_exception(): [ 565.544150] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 565.544150] env[63088]: self.force_reraise() [ 565.544150] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 565.544150] env[63088]: raise self.value [ 565.544150] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 565.544150] env[63088]: updated_port = self._update_port( [ 565.544150] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 565.544150] env[63088]: _ensure_no_port_binding_failure(port) [ 565.544150] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 565.544150] env[63088]: raise exception.PortBindingFailed(port_id=port['id']) [ 565.544994] env[63088]: nova.exception.PortBindingFailed: Binding failed for port 06dbc948-3bcb-4bb1-9083-7a5a312da444, please check neutron logs for more information. [ 565.544994] env[63088]: Removing descriptor: 16 [ 565.544994] env[63088]: ERROR nova.compute.manager [None req-61dd383c-8cef-4a29-a464-69dfd03cb5ce tempest-ServerExternalEventsTest-515168927 tempest-ServerExternalEventsTest-515168927-project-member] [instance: 173fccf1-f258-4f0c-a968-466a27406c55] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 06dbc948-3bcb-4bb1-9083-7a5a312da444, please check neutron logs for more information. [ 565.544994] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] Traceback (most recent call last): [ 565.544994] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 565.544994] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] yield resources [ 565.544994] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 565.544994] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] self.driver.spawn(context, instance, image_meta, [ 565.544994] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 565.544994] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] self._vmops.spawn(context, instance, image_meta, injected_files, [ 565.544994] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 565.544994] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] vm_ref = self.build_virtual_machine(instance, [ 565.545342] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 565.545342] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] vif_infos = vmwarevif.get_vif_info(self._session, [ 565.545342] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 565.545342] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] for vif in network_info: [ 565.545342] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 565.545342] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] return self._sync_wrapper(fn, *args, **kwargs) [ 565.545342] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 565.545342] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] self.wait() [ 565.545342] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 565.545342] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] self[:] = self._gt.wait() [ 565.545342] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 565.545342] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] return self._exit_event.wait() [ 565.545342] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 565.545761] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] result = hub.switch() [ 565.545761] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 565.545761] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] return self.greenlet.switch() [ 565.545761] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 565.545761] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] result = function(*args, **kwargs) [ 565.545761] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 565.545761] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] return func(*args, **kwargs) [ 565.545761] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 565.545761] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] raise e [ 565.545761] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 565.545761] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] nwinfo = self.network_api.allocate_for_instance( [ 565.545761] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 565.545761] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] created_port_ids = self._update_ports_for_instance( [ 565.546193] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 565.546193] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] with excutils.save_and_reraise_exception(): [ 565.546193] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 565.546193] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] self.force_reraise() [ 565.546193] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 565.546193] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] raise self.value [ 565.546193] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 565.546193] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] updated_port = self._update_port( [ 565.546193] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 565.546193] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] _ensure_no_port_binding_failure(port) [ 565.546193] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 565.546193] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] raise exception.PortBindingFailed(port_id=port['id']) [ 565.546572] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] nova.exception.PortBindingFailed: Binding failed for port 06dbc948-3bcb-4bb1-9083-7a5a312da444, please check neutron logs for more information. [ 565.546572] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] [ 565.546572] env[63088]: INFO nova.compute.manager [None req-61dd383c-8cef-4a29-a464-69dfd03cb5ce tempest-ServerExternalEventsTest-515168927 tempest-ServerExternalEventsTest-515168927-project-member] [instance: 173fccf1-f258-4f0c-a968-466a27406c55] Terminating instance [ 565.549296] env[63088]: DEBUG oslo_concurrency.lockutils [None req-61dd383c-8cef-4a29-a464-69dfd03cb5ce tempest-ServerExternalEventsTest-515168927 tempest-ServerExternalEventsTest-515168927-project-member] Acquiring lock "refresh_cache-173fccf1-f258-4f0c-a968-466a27406c55" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 565.549296] env[63088]: DEBUG oslo_concurrency.lockutils [None req-61dd383c-8cef-4a29-a464-69dfd03cb5ce tempest-ServerExternalEventsTest-515168927 tempest-ServerExternalEventsTest-515168927-project-member] Acquired lock "refresh_cache-173fccf1-f258-4f0c-a968-466a27406c55" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 565.549443] env[63088]: DEBUG nova.network.neutron [None req-61dd383c-8cef-4a29-a464-69dfd03cb5ce tempest-ServerExternalEventsTest-515168927 tempest-ServerExternalEventsTest-515168927-project-member] [instance: 173fccf1-f258-4f0c-a968-466a27406c55] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 565.551476] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e248a83f-0cca-462e-847e-5a95323c1892 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.568716] env[63088]: DEBUG nova.compute.provider_tree [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Updating inventory in ProviderTree for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 565.823193] env[63088]: DEBUG oslo_vmware.api [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Task: {'id': task-1284769, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066829} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 565.823433] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Extended root virtual disk {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 565.824324] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f766b27-b5db-44e8-880d-073697ca19b0 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.857078] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Reconfiguring VM instance instance-00000005 to attach disk [datastore1] e8cfbce4-cfb2-463e-9f38-78a058dab6e1/e8cfbce4-cfb2-463e-9f38-78a058dab6e1.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 565.857707] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cc5bc9c9-ac86-4804-b2ec-4881ba9cf15a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.883017] env[63088]: DEBUG oslo_vmware.api [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Waiting for the task: (returnval){ [ 565.883017] env[63088]: value = "task-1284770" [ 565.883017] env[63088]: _type = "Task" [ 565.883017] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 565.890779] env[63088]: DEBUG oslo_vmware.api [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Task: {'id': task-1284770, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 566.098570] env[63088]: ERROR nova.scheduler.client.report [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] [req-10330e01-ec93-4bbb-b83b-8b2f4fdf8e17] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 6eae54a9-8831-40eb-bf54-4bc60d346b02. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-10330e01-ec93-4bbb-b83b-8b2f4fdf8e17"}]} [ 566.100820] env[63088]: DEBUG nova.network.neutron [None req-61dd383c-8cef-4a29-a464-69dfd03cb5ce tempest-ServerExternalEventsTest-515168927 tempest-ServerExternalEventsTest-515168927-project-member] [instance: 173fccf1-f258-4f0c-a968-466a27406c55] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 566.115877] env[63088]: DEBUG nova.scheduler.client.report [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Refreshing inventories for resource provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 566.135708] env[63088]: DEBUG nova.scheduler.client.report [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Updating ProviderTree inventory for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 566.135708] env[63088]: DEBUG nova.compute.provider_tree [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Updating inventory in ProviderTree for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 566.152227] env[63088]: DEBUG nova.scheduler.client.report [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Refreshing aggregate associations for resource provider 6eae54a9-8831-40eb-bf54-4bc60d346b02, aggregates: None {{(pid=63088) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 566.173288] env[63088]: DEBUG nova.scheduler.client.report [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Refreshing trait associations for resource provider 6eae54a9-8831-40eb-bf54-4bc60d346b02, traits: COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO {{(pid=63088) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 566.280015] env[63088]: DEBUG nova.network.neutron [None req-61dd383c-8cef-4a29-a464-69dfd03cb5ce tempest-ServerExternalEventsTest-515168927 tempest-ServerExternalEventsTest-515168927-project-member] [instance: 173fccf1-f258-4f0c-a968-466a27406c55] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 566.398090] env[63088]: DEBUG oslo_vmware.api [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Task: {'id': task-1284770, 'name': ReconfigVM_Task, 'duration_secs': 0.297804} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 566.398756] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Reconfigured VM instance instance-00000005 to attach disk [datastore1] e8cfbce4-cfb2-463e-9f38-78a058dab6e1/e8cfbce4-cfb2-463e-9f38-78a058dab6e1.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 566.399507] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ccb40f8e-a616-481c-8a53-dbf17cdee3f1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.408060] env[63088]: DEBUG oslo_vmware.api [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Waiting for the task: (returnval){ [ 566.408060] env[63088]: value = "task-1284771" [ 566.408060] env[63088]: _type = "Task" [ 566.408060] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 566.421925] env[63088]: DEBUG oslo_vmware.api [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Task: {'id': task-1284771, 'name': Rename_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 566.425712] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a82adb6-04c8-4987-9421-9e2d2f97ffd6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.433310] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef7f2870-30e8-4e28-b22e-7a91c0fa6500 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.471306] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2a42e33-e114-4b53-b411-0f6f49a50323 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.477639] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bd71c985-753f-4eb9-b297-7d458ea844aa tempest-ImagesOneServerNegativeTestJSON-1975365234 tempest-ImagesOneServerNegativeTestJSON-1975365234-project-member] Acquiring lock "ca92ddb4-cac8-450a-9b7e-d09ad082354a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 566.477639] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bd71c985-753f-4eb9-b297-7d458ea844aa tempest-ImagesOneServerNegativeTestJSON-1975365234 tempest-ImagesOneServerNegativeTestJSON-1975365234-project-member] Lock "ca92ddb4-cac8-450a-9b7e-d09ad082354a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 566.483617] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ac06851-9be1-4e4e-b85f-f3ccc15aea6e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.500244] env[63088]: DEBUG nova.compute.provider_tree [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Updating inventory in ProviderTree for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 566.787377] env[63088]: DEBUG oslo_concurrency.lockutils [None req-61dd383c-8cef-4a29-a464-69dfd03cb5ce tempest-ServerExternalEventsTest-515168927 tempest-ServerExternalEventsTest-515168927-project-member] Releasing lock "refresh_cache-173fccf1-f258-4f0c-a968-466a27406c55" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 566.787610] env[63088]: DEBUG nova.compute.manager [None req-61dd383c-8cef-4a29-a464-69dfd03cb5ce tempest-ServerExternalEventsTest-515168927 tempest-ServerExternalEventsTest-515168927-project-member] [instance: 173fccf1-f258-4f0c-a968-466a27406c55] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 566.787836] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-61dd383c-8cef-4a29-a464-69dfd03cb5ce tempest-ServerExternalEventsTest-515168927 tempest-ServerExternalEventsTest-515168927-project-member] [instance: 173fccf1-f258-4f0c-a968-466a27406c55] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 566.788162] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a0ea5f29-e077-45f6-a374-bcca943be5b3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.800911] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf870b99-5abd-4552-be0e-d2a88d5afaf6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.825051] env[63088]: WARNING nova.virt.vmwareapi.vmops [None req-61dd383c-8cef-4a29-a464-69dfd03cb5ce tempest-ServerExternalEventsTest-515168927 tempest-ServerExternalEventsTest-515168927-project-member] [instance: 173fccf1-f258-4f0c-a968-466a27406c55] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 173fccf1-f258-4f0c-a968-466a27406c55 could not be found. [ 566.825051] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-61dd383c-8cef-4a29-a464-69dfd03cb5ce tempest-ServerExternalEventsTest-515168927 tempest-ServerExternalEventsTest-515168927-project-member] [instance: 173fccf1-f258-4f0c-a968-466a27406c55] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 566.825182] env[63088]: INFO nova.compute.manager [None req-61dd383c-8cef-4a29-a464-69dfd03cb5ce tempest-ServerExternalEventsTest-515168927 tempest-ServerExternalEventsTest-515168927-project-member] [instance: 173fccf1-f258-4f0c-a968-466a27406c55] Took 0.04 seconds to destroy the instance on the hypervisor. [ 566.825360] env[63088]: DEBUG oslo.service.loopingcall [None req-61dd383c-8cef-4a29-a464-69dfd03cb5ce tempest-ServerExternalEventsTest-515168927 tempest-ServerExternalEventsTest-515168927-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 566.825559] env[63088]: DEBUG nova.compute.manager [-] [instance: 173fccf1-f258-4f0c-a968-466a27406c55] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 566.825670] env[63088]: DEBUG nova.network.neutron [-] [instance: 173fccf1-f258-4f0c-a968-466a27406c55] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 566.860058] env[63088]: DEBUG nova.network.neutron [-] [instance: 173fccf1-f258-4f0c-a968-466a27406c55] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 566.924080] env[63088]: DEBUG oslo_vmware.api [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Task: {'id': task-1284771, 'name': Rename_Task, 'duration_secs': 0.176124} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 566.924080] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 566.924330] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3d3b4e86-d8ca-445f-8330-24ae45138a30 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.931809] env[63088]: DEBUG oslo_vmware.api [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Waiting for the task: (returnval){ [ 566.931809] env[63088]: value = "task-1284772" [ 566.931809] env[63088]: _type = "Task" [ 566.931809] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 566.942020] env[63088]: DEBUG oslo_vmware.api [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Task: {'id': task-1284772, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 567.041791] env[63088]: DEBUG nova.scheduler.client.report [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Updated inventory for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 with generation 18 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 567.041901] env[63088]: DEBUG nova.compute.provider_tree [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Updating resource provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 generation from 18 to 19 during operation: update_inventory {{(pid=63088) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 567.042087] env[63088]: DEBUG nova.compute.provider_tree [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Updating inventory in ProviderTree for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 567.362969] env[63088]: DEBUG nova.network.neutron [-] [instance: 173fccf1-f258-4f0c-a968-466a27406c55] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 567.443470] env[63088]: DEBUG oslo_vmware.api [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Task: {'id': task-1284772, 'name': PowerOnVM_Task, 'duration_secs': 0.434469} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 567.444752] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 567.445026] env[63088]: INFO nova.compute.manager [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Took 9.15 seconds to spawn the instance on the hypervisor. [ 567.445329] env[63088]: DEBUG nova.compute.manager [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 567.448106] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b83e7013-bbf8-470d-8e20-21c771bc0614 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.552016] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 4.242s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 567.552618] env[63088]: DEBUG nova.compute.manager [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] [instance: c4777721-3f65-455f-9973-c1ed0732de34] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 567.555588] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6e35ce22-cf37-483a-8093-9574de222716 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.886s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 567.560258] env[63088]: INFO nova.compute.claims [None req-6e35ce22-cf37-483a-8093-9574de222716 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 567.868498] env[63088]: INFO nova.compute.manager [-] [instance: 173fccf1-f258-4f0c-a968-466a27406c55] Took 1.04 seconds to deallocate network for instance. [ 567.872711] env[63088]: DEBUG nova.compute.claims [None req-61dd383c-8cef-4a29-a464-69dfd03cb5ce tempest-ServerExternalEventsTest-515168927 tempest-ServerExternalEventsTest-515168927-project-member] [instance: 173fccf1-f258-4f0c-a968-466a27406c55] Aborting claim: {{(pid=63088) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 567.873020] env[63088]: DEBUG oslo_concurrency.lockutils [None req-61dd383c-8cef-4a29-a464-69dfd03cb5ce tempest-ServerExternalEventsTest-515168927 tempest-ServerExternalEventsTest-515168927-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 567.971534] env[63088]: INFO nova.compute.manager [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Took 21.17 seconds to build instance. [ 568.068685] env[63088]: DEBUG nova.compute.utils [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 568.068685] env[63088]: DEBUG nova.compute.manager [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] [instance: c4777721-3f65-455f-9973-c1ed0732de34] Not allocating networking since 'none' was specified. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 568.233724] env[63088]: ERROR nova.compute.manager [None req-6568dcbe-ebf7-46f4-8cbf-a6cec6f2c264 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 0c9febcc-5137-47a2-a0b9-df090f78c2e7, please check neutron logs for more information. [ 568.233724] env[63088]: ERROR nova.compute.manager Traceback (most recent call last): [ 568.233724] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 568.233724] env[63088]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 568.233724] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 568.233724] env[63088]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 568.233724] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 568.233724] env[63088]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 568.233724] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 568.233724] env[63088]: ERROR nova.compute.manager self.force_reraise() [ 568.233724] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 568.233724] env[63088]: ERROR nova.compute.manager raise self.value [ 568.233724] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 568.233724] env[63088]: ERROR nova.compute.manager updated_port = self._update_port( [ 568.233724] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 568.233724] env[63088]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 568.234642] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 568.234642] env[63088]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 568.234642] env[63088]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 0c9febcc-5137-47a2-a0b9-df090f78c2e7, please check neutron logs for more information. [ 568.234642] env[63088]: ERROR nova.compute.manager [ 568.234642] env[63088]: Traceback (most recent call last): [ 568.234642] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 568.234642] env[63088]: listener.cb(fileno) [ 568.234642] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 568.234642] env[63088]: result = function(*args, **kwargs) [ 568.234642] env[63088]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 568.234642] env[63088]: return func(*args, **kwargs) [ 568.234642] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 568.234642] env[63088]: raise e [ 568.234642] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 568.234642] env[63088]: nwinfo = self.network_api.allocate_for_instance( [ 568.234642] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 568.234642] env[63088]: created_port_ids = self._update_ports_for_instance( [ 568.234642] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 568.234642] env[63088]: with excutils.save_and_reraise_exception(): [ 568.234642] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 568.234642] env[63088]: self.force_reraise() [ 568.234642] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 568.234642] env[63088]: raise self.value [ 568.234642] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 568.234642] env[63088]: updated_port = self._update_port( [ 568.234642] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 568.234642] env[63088]: _ensure_no_port_binding_failure(port) [ 568.234642] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 568.234642] env[63088]: raise exception.PortBindingFailed(port_id=port['id']) [ 568.235365] env[63088]: nova.exception.PortBindingFailed: Binding failed for port 0c9febcc-5137-47a2-a0b9-df090f78c2e7, please check neutron logs for more information. [ 568.235365] env[63088]: Removing descriptor: 15 [ 568.235365] env[63088]: ERROR nova.compute.manager [None req-6568dcbe-ebf7-46f4-8cbf-a6cec6f2c264 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] [instance: 4d5293a0-6945-4731-afb6-996a142c8447] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 0c9febcc-5137-47a2-a0b9-df090f78c2e7, please check neutron logs for more information. [ 568.235365] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] Traceback (most recent call last): [ 568.235365] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 568.235365] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] yield resources [ 568.235365] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 568.235365] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] self.driver.spawn(context, instance, image_meta, [ 568.235365] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 568.235365] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] self._vmops.spawn(context, instance, image_meta, injected_files, [ 568.235365] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 568.235365] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] vm_ref = self.build_virtual_machine(instance, [ 568.235806] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 568.235806] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] vif_infos = vmwarevif.get_vif_info(self._session, [ 568.235806] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 568.235806] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] for vif in network_info: [ 568.235806] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 568.235806] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] return self._sync_wrapper(fn, *args, **kwargs) [ 568.235806] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 568.235806] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] self.wait() [ 568.235806] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 568.235806] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] self[:] = self._gt.wait() [ 568.235806] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 568.235806] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] return self._exit_event.wait() [ 568.235806] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 568.236296] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] result = hub.switch() [ 568.236296] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 568.236296] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] return self.greenlet.switch() [ 568.236296] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 568.236296] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] result = function(*args, **kwargs) [ 568.236296] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 568.236296] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] return func(*args, **kwargs) [ 568.236296] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 568.236296] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] raise e [ 568.236296] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 568.236296] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] nwinfo = self.network_api.allocate_for_instance( [ 568.236296] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 568.236296] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] created_port_ids = self._update_ports_for_instance( [ 568.239579] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 568.239579] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] with excutils.save_and_reraise_exception(): [ 568.239579] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 568.239579] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] self.force_reraise() [ 568.239579] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 568.239579] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] raise self.value [ 568.239579] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 568.239579] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] updated_port = self._update_port( [ 568.239579] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 568.239579] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] _ensure_no_port_binding_failure(port) [ 568.239579] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 568.239579] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] raise exception.PortBindingFailed(port_id=port['id']) [ 568.239920] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] nova.exception.PortBindingFailed: Binding failed for port 0c9febcc-5137-47a2-a0b9-df090f78c2e7, please check neutron logs for more information. [ 568.239920] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] [ 568.239920] env[63088]: INFO nova.compute.manager [None req-6568dcbe-ebf7-46f4-8cbf-a6cec6f2c264 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] [instance: 4d5293a0-6945-4731-afb6-996a142c8447] Terminating instance [ 568.239920] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6568dcbe-ebf7-46f4-8cbf-a6cec6f2c264 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] Acquiring lock "refresh_cache-4d5293a0-6945-4731-afb6-996a142c8447" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 568.239920] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6568dcbe-ebf7-46f4-8cbf-a6cec6f2c264 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] Acquired lock "refresh_cache-4d5293a0-6945-4731-afb6-996a142c8447" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 568.239920] env[63088]: DEBUG nova.network.neutron [None req-6568dcbe-ebf7-46f4-8cbf-a6cec6f2c264 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] [instance: 4d5293a0-6945-4731-afb6-996a142c8447] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 568.474377] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c0a3780e-3e41-452e-9e4f-3f16668512e2 tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Lock "e8cfbce4-cfb2-463e-9f38-78a058dab6e1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.678s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 568.503110] env[63088]: DEBUG nova.compute.manager [req-79fd4314-120f-4300-84ed-fb89c7e97347 req-d2fe214c-7b0d-4dd7-8033-e482fc14f3eb service nova] [instance: 173fccf1-f258-4f0c-a968-466a27406c55] Received event network-changed-06dbc948-3bcb-4bb1-9083-7a5a312da444 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 568.503312] env[63088]: DEBUG nova.compute.manager [req-79fd4314-120f-4300-84ed-fb89c7e97347 req-d2fe214c-7b0d-4dd7-8033-e482fc14f3eb service nova] [instance: 173fccf1-f258-4f0c-a968-466a27406c55] Refreshing instance network info cache due to event network-changed-06dbc948-3bcb-4bb1-9083-7a5a312da444. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 568.503509] env[63088]: DEBUG oslo_concurrency.lockutils [req-79fd4314-120f-4300-84ed-fb89c7e97347 req-d2fe214c-7b0d-4dd7-8033-e482fc14f3eb service nova] Acquiring lock "refresh_cache-173fccf1-f258-4f0c-a968-466a27406c55" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 568.503667] env[63088]: DEBUG oslo_concurrency.lockutils [req-79fd4314-120f-4300-84ed-fb89c7e97347 req-d2fe214c-7b0d-4dd7-8033-e482fc14f3eb service nova] Acquired lock "refresh_cache-173fccf1-f258-4f0c-a968-466a27406c55" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 568.503819] env[63088]: DEBUG nova.network.neutron [req-79fd4314-120f-4300-84ed-fb89c7e97347 req-d2fe214c-7b0d-4dd7-8033-e482fc14f3eb service nova] [instance: 173fccf1-f258-4f0c-a968-466a27406c55] Refreshing network info cache for port 06dbc948-3bcb-4bb1-9083-7a5a312da444 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 568.573115] env[63088]: DEBUG nova.compute.manager [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] [instance: c4777721-3f65-455f-9973-c1ed0732de34] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 568.776753] env[63088]: DEBUG nova.network.neutron [None req-6568dcbe-ebf7-46f4-8cbf-a6cec6f2c264 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] [instance: 4d5293a0-6945-4731-afb6-996a142c8447] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 568.797529] env[63088]: DEBUG oslo_concurrency.lockutils [None req-83cac04a-fc75-40e5-8cdf-832028033c8d tempest-ServersWithSpecificFlavorTestJSON-501634894 tempest-ServersWithSpecificFlavorTestJSON-501634894-project-member] Acquiring lock "d6c76851-7059-40a4-a39b-e625c99dfcd9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 568.797812] env[63088]: DEBUG oslo_concurrency.lockutils [None req-83cac04a-fc75-40e5-8cdf-832028033c8d tempest-ServersWithSpecificFlavorTestJSON-501634894 tempest-ServersWithSpecificFlavorTestJSON-501634894-project-member] Lock "d6c76851-7059-40a4-a39b-e625c99dfcd9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 568.881182] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52eaeec9-7644-47a5-90e1-596589e19252 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.889782] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-591f4524-2382-40f5-a061-4e28d40a5bd6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.926136] env[63088]: ERROR nova.compute.manager [None req-bc610745-9520-42a6-9fd1-7e4c9c63b77a tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 010a5358-1268-49c2-b2b4-9525552d6fdc, please check neutron logs for more information. [ 568.926136] env[63088]: ERROR nova.compute.manager Traceback (most recent call last): [ 568.926136] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 568.926136] env[63088]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 568.926136] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 568.926136] env[63088]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 568.926136] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 568.926136] env[63088]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 568.926136] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 568.926136] env[63088]: ERROR nova.compute.manager self.force_reraise() [ 568.926136] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 568.926136] env[63088]: ERROR nova.compute.manager raise self.value [ 568.926136] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 568.926136] env[63088]: ERROR nova.compute.manager updated_port = self._update_port( [ 568.926136] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 568.926136] env[63088]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 568.927623] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 568.927623] env[63088]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 568.927623] env[63088]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 010a5358-1268-49c2-b2b4-9525552d6fdc, please check neutron logs for more information. [ 568.927623] env[63088]: ERROR nova.compute.manager [ 568.927623] env[63088]: Traceback (most recent call last): [ 568.927623] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 568.927623] env[63088]: listener.cb(fileno) [ 568.927623] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 568.927623] env[63088]: result = function(*args, **kwargs) [ 568.927623] env[63088]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 568.927623] env[63088]: return func(*args, **kwargs) [ 568.927623] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 568.927623] env[63088]: raise e [ 568.927623] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 568.927623] env[63088]: nwinfo = self.network_api.allocate_for_instance( [ 568.927623] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 568.927623] env[63088]: created_port_ids = self._update_ports_for_instance( [ 568.927623] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 568.927623] env[63088]: with excutils.save_and_reraise_exception(): [ 568.927623] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 568.927623] env[63088]: self.force_reraise() [ 568.927623] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 568.927623] env[63088]: raise self.value [ 568.927623] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 568.927623] env[63088]: updated_port = self._update_port( [ 568.927623] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 568.927623] env[63088]: _ensure_no_port_binding_failure(port) [ 568.927623] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 568.927623] env[63088]: raise exception.PortBindingFailed(port_id=port['id']) [ 568.928672] env[63088]: nova.exception.PortBindingFailed: Binding failed for port 010a5358-1268-49c2-b2b4-9525552d6fdc, please check neutron logs for more information. [ 568.928672] env[63088]: Removing descriptor: 17 [ 568.928672] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca44337d-383f-4bc2-9d57-ac1153432726 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.930285] env[63088]: ERROR nova.compute.manager [None req-bc610745-9520-42a6-9fd1-7e4c9c63b77a tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 010a5358-1268-49c2-b2b4-9525552d6fdc, please check neutron logs for more information. [ 568.930285] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] Traceback (most recent call last): [ 568.930285] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 568.930285] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] yield resources [ 568.930285] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 568.930285] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] self.driver.spawn(context, instance, image_meta, [ 568.930285] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 568.930285] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] self._vmops.spawn(context, instance, image_meta, injected_files, [ 568.930285] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 568.930285] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] vm_ref = self.build_virtual_machine(instance, [ 568.930285] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 568.930760] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] vif_infos = vmwarevif.get_vif_info(self._session, [ 568.930760] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 568.930760] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] for vif in network_info: [ 568.930760] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 568.930760] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] return self._sync_wrapper(fn, *args, **kwargs) [ 568.930760] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 568.930760] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] self.wait() [ 568.930760] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 568.930760] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] self[:] = self._gt.wait() [ 568.930760] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 568.930760] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] return self._exit_event.wait() [ 568.930760] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 568.930760] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] result = hub.switch() [ 568.931145] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 568.931145] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] return self.greenlet.switch() [ 568.931145] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 568.931145] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] result = function(*args, **kwargs) [ 568.931145] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 568.931145] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] return func(*args, **kwargs) [ 568.931145] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 568.931145] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] raise e [ 568.931145] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 568.931145] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] nwinfo = self.network_api.allocate_for_instance( [ 568.931145] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 568.931145] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] created_port_ids = self._update_ports_for_instance( [ 568.931145] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 568.931519] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] with excutils.save_and_reraise_exception(): [ 568.931519] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 568.931519] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] self.force_reraise() [ 568.931519] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 568.931519] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] raise self.value [ 568.931519] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 568.931519] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] updated_port = self._update_port( [ 568.931519] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 568.931519] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] _ensure_no_port_binding_failure(port) [ 568.931519] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 568.931519] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] raise exception.PortBindingFailed(port_id=port['id']) [ 568.931519] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] nova.exception.PortBindingFailed: Binding failed for port 010a5358-1268-49c2-b2b4-9525552d6fdc, please check neutron logs for more information. [ 568.931519] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] [ 568.931932] env[63088]: INFO nova.compute.manager [None req-bc610745-9520-42a6-9fd1-7e4c9c63b77a tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] Terminating instance [ 568.933158] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bc610745-9520-42a6-9fd1-7e4c9c63b77a tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] Acquiring lock "refresh_cache-5c3f72e3-d4f4-4c34-92a3-075536b85c99" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 568.933301] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bc610745-9520-42a6-9fd1-7e4c9c63b77a tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] Acquired lock "refresh_cache-5c3f72e3-d4f4-4c34-92a3-075536b85c99" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 568.933470] env[63088]: DEBUG nova.network.neutron [None req-bc610745-9520-42a6-9fd1-7e4c9c63b77a tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 568.940227] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c1ca368-e6a3-4c70-9c84-51bc5ae0c772 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.958963] env[63088]: DEBUG nova.compute.provider_tree [None req-6e35ce22-cf37-483a-8093-9574de222716 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 568.960885] env[63088]: DEBUG nova.network.neutron [None req-6568dcbe-ebf7-46f4-8cbf-a6cec6f2c264 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] [instance: 4d5293a0-6945-4731-afb6-996a142c8447] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 568.976480] env[63088]: DEBUG nova.compute.manager [None req-d1272b04-ee5c-4e9e-b0bf-6ca9d0629aae tempest-VolumesAssistedSnapshotsTest-1004656420 tempest-VolumesAssistedSnapshotsTest-1004656420-project-member] [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 569.061389] env[63088]: DEBUG nova.network.neutron [req-79fd4314-120f-4300-84ed-fb89c7e97347 req-d2fe214c-7b0d-4dd7-8033-e482fc14f3eb service nova] [instance: 173fccf1-f258-4f0c-a968-466a27406c55] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 569.195199] env[63088]: DEBUG nova.network.neutron [req-79fd4314-120f-4300-84ed-fb89c7e97347 req-d2fe214c-7b0d-4dd7-8033-e482fc14f3eb service nova] [instance: 173fccf1-f258-4f0c-a968-466a27406c55] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 569.466315] env[63088]: DEBUG nova.scheduler.client.report [None req-6e35ce22-cf37-483a-8093-9574de222716 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 569.469074] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6568dcbe-ebf7-46f4-8cbf-a6cec6f2c264 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] Releasing lock "refresh_cache-4d5293a0-6945-4731-afb6-996a142c8447" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 569.474671] env[63088]: DEBUG nova.compute.manager [None req-6568dcbe-ebf7-46f4-8cbf-a6cec6f2c264 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] [instance: 4d5293a0-6945-4731-afb6-996a142c8447] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 569.474671] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-6568dcbe-ebf7-46f4-8cbf-a6cec6f2c264 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] [instance: 4d5293a0-6945-4731-afb6-996a142c8447] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 569.477017] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e2ff5f55-3cc0-47dc-bcbe-ac2db3049404 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.489528] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-948df63c-eff0-4789-a5a2-cc4b75a7d16c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.504419] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d1272b04-ee5c-4e9e-b0bf-6ca9d0629aae tempest-VolumesAssistedSnapshotsTest-1004656420 tempest-VolumesAssistedSnapshotsTest-1004656420-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 569.518777] env[63088]: WARNING nova.virt.vmwareapi.vmops [None req-6568dcbe-ebf7-46f4-8cbf-a6cec6f2c264 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] [instance: 4d5293a0-6945-4731-afb6-996a142c8447] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4d5293a0-6945-4731-afb6-996a142c8447 could not be found. [ 569.519179] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-6568dcbe-ebf7-46f4-8cbf-a6cec6f2c264 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] [instance: 4d5293a0-6945-4731-afb6-996a142c8447] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 569.519528] env[63088]: INFO nova.compute.manager [None req-6568dcbe-ebf7-46f4-8cbf-a6cec6f2c264 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] [instance: 4d5293a0-6945-4731-afb6-996a142c8447] Took 0.05 seconds to destroy the instance on the hypervisor. [ 569.519892] env[63088]: DEBUG oslo.service.loopingcall [None req-6568dcbe-ebf7-46f4-8cbf-a6cec6f2c264 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 569.523020] env[63088]: DEBUG nova.network.neutron [None req-bc610745-9520-42a6-9fd1-7e4c9c63b77a tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 569.523020] env[63088]: DEBUG nova.compute.manager [-] [instance: 4d5293a0-6945-4731-afb6-996a142c8447] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 569.523020] env[63088]: DEBUG nova.network.neutron [-] [instance: 4d5293a0-6945-4731-afb6-996a142c8447] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 569.553371] env[63088]: DEBUG nova.network.neutron [-] [instance: 4d5293a0-6945-4731-afb6-996a142c8447] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 569.585685] env[63088]: DEBUG nova.compute.manager [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] [instance: c4777721-3f65-455f-9973-c1ed0732de34] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 569.615452] env[63088]: DEBUG nova.virt.hardware [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 569.615744] env[63088]: DEBUG nova.virt.hardware [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 569.616033] env[63088]: DEBUG nova.virt.hardware [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 569.616146] env[63088]: DEBUG nova.virt.hardware [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 569.616234] env[63088]: DEBUG nova.virt.hardware [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 569.616375] env[63088]: DEBUG nova.virt.hardware [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 569.616880] env[63088]: DEBUG nova.virt.hardware [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 569.616880] env[63088]: DEBUG nova.virt.hardware [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 569.617018] env[63088]: DEBUG nova.virt.hardware [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 569.617090] env[63088]: DEBUG nova.virt.hardware [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 569.617262] env[63088]: DEBUG nova.virt.hardware [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 569.618206] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9084f740-0e18-4b3a-acbb-33e65aa24745 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.628317] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc6dfe11-02d3-4fc0-89b1-46d668badc51 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.645025] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] [instance: c4777721-3f65-455f-9973-c1ed0732de34] Instance VIF info [] {{(pid=63088) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 569.651527] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Creating folder: Project (de057d7a4734400589e8712607249e42). Parent ref: group-v275816. {{(pid=63088) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 569.651587] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0db38943-217d-47cb-b0b3-dd26b33c32bd {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.661128] env[63088]: DEBUG nova.compute.manager [req-c15b05cf-c77d-446a-a7b3-25e7c22d9a3f req-1111477b-a942-428a-a716-64128cbda1b3 service nova] [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] Received event network-changed-010a5358-1268-49c2-b2b4-9525552d6fdc {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 569.663402] env[63088]: DEBUG nova.compute.manager [req-c15b05cf-c77d-446a-a7b3-25e7c22d9a3f req-1111477b-a942-428a-a716-64128cbda1b3 service nova] [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] Refreshing instance network info cache due to event network-changed-010a5358-1268-49c2-b2b4-9525552d6fdc. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 569.663402] env[63088]: DEBUG oslo_concurrency.lockutils [req-c15b05cf-c77d-446a-a7b3-25e7c22d9a3f req-1111477b-a942-428a-a716-64128cbda1b3 service nova] Acquiring lock "refresh_cache-5c3f72e3-d4f4-4c34-92a3-075536b85c99" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 569.663741] env[63088]: INFO nova.virt.vmwareapi.vm_util [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Created folder: Project (de057d7a4734400589e8712607249e42) in parent group-v275816. [ 569.663741] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Creating folder: Instances. Parent ref: group-v275820. {{(pid=63088) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 569.663949] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c23e2c94-7404-4bae-8f84-5fae7cbc941b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.676829] env[63088]: INFO nova.virt.vmwareapi.vm_util [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Created folder: Instances in parent group-v275820. [ 569.676829] env[63088]: DEBUG oslo.service.loopingcall [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 569.676829] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c4777721-3f65-455f-9973-c1ed0732de34] Creating VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 569.677021] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ee344dbd-7a26-4485-9504-dda23730815f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.698021] env[63088]: DEBUG oslo_concurrency.lockutils [req-79fd4314-120f-4300-84ed-fb89c7e97347 req-d2fe214c-7b0d-4dd7-8033-e482fc14f3eb service nova] Releasing lock "refresh_cache-173fccf1-f258-4f0c-a968-466a27406c55" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 569.700508] env[63088]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 569.700508] env[63088]: value = "task-1284775" [ 569.700508] env[63088]: _type = "Task" [ 569.700508] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 569.713573] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1284775, 'name': CreateVM_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 569.716033] env[63088]: DEBUG nova.network.neutron [None req-bc610745-9520-42a6-9fd1-7e4c9c63b77a tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 569.980344] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6e35ce22-cf37-483a-8093-9574de222716 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.425s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 569.980872] env[63088]: DEBUG nova.compute.manager [None req-6e35ce22-cf37-483a-8093-9574de222716 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 569.986679] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 16.345s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 569.986679] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 569.986784] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63088) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 569.986997] env[63088]: DEBUG oslo_concurrency.lockutils [None req-96b6b7c1-255c-445f-b8db-9f4849d061fb tempest-ImagesNegativeTestJSON-1434405362 tempest-ImagesNegativeTestJSON-1434405362-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.839s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 569.990983] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd11608b-2d0e-4f44-b4b5-ddf13a38eca2 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.002369] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6617a3c-1466-4bbc-95b5-f162235cd939 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.016796] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74ab13f5-9f71-4232-a6bc-4e9c1424857d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.023843] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95fca055-e934-4924-8b29-81eeacffa73f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.062438] env[63088]: DEBUG nova.network.neutron [-] [instance: 4d5293a0-6945-4731-afb6-996a142c8447] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 570.063892] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181496MB free_disk=140GB free_vcpus=48 pci_devices=None {{(pid=63088) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 570.064046] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 570.108065] env[63088]: DEBUG oslo_concurrency.lockutils [None req-acfde423-bb82-457b-a17c-3c463abca710 tempest-ServersTestFqdnHostnames-1568147653 tempest-ServersTestFqdnHostnames-1568147653-project-member] Acquiring lock "91a09343-c2d7-4f5d-ae0a-4a123af92fba" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 570.109303] env[63088]: DEBUG oslo_concurrency.lockutils [None req-acfde423-bb82-457b-a17c-3c463abca710 tempest-ServersTestFqdnHostnames-1568147653 tempest-ServersTestFqdnHostnames-1568147653-project-member] Lock "91a09343-c2d7-4f5d-ae0a-4a123af92fba" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 570.219871] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bc610745-9520-42a6-9fd1-7e4c9c63b77a tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] Releasing lock "refresh_cache-5c3f72e3-d4f4-4c34-92a3-075536b85c99" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 570.220433] env[63088]: DEBUG nova.compute.manager [None req-bc610745-9520-42a6-9fd1-7e4c9c63b77a tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 570.220546] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-bc610745-9520-42a6-9fd1-7e4c9c63b77a tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 570.220773] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1284775, 'name': CreateVM_Task} progress is 99%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 570.220995] env[63088]: DEBUG oslo_concurrency.lockutils [req-c15b05cf-c77d-446a-a7b3-25e7c22d9a3f req-1111477b-a942-428a-a716-64128cbda1b3 service nova] Acquired lock "refresh_cache-5c3f72e3-d4f4-4c34-92a3-075536b85c99" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 570.221171] env[63088]: DEBUG nova.network.neutron [req-c15b05cf-c77d-446a-a7b3-25e7c22d9a3f req-1111477b-a942-428a-a716-64128cbda1b3 service nova] [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] Refreshing network info cache for port 010a5358-1268-49c2-b2b4-9525552d6fdc {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 570.223680] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f450ae58-1e7c-4b12-95b1-4ed890d1e4b0 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.233510] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d463afd5-19ef-4dd8-8627-fd9f13388639 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.258300] env[63088]: WARNING nova.virt.vmwareapi.vmops [None req-bc610745-9520-42a6-9fd1-7e4c9c63b77a tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5c3f72e3-d4f4-4c34-92a3-075536b85c99 could not be found. [ 570.258514] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-bc610745-9520-42a6-9fd1-7e4c9c63b77a tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 570.258705] env[63088]: INFO nova.compute.manager [None req-bc610745-9520-42a6-9fd1-7e4c9c63b77a tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] Took 0.04 seconds to destroy the instance on the hypervisor. [ 570.259023] env[63088]: DEBUG oslo.service.loopingcall [None req-bc610745-9520-42a6-9fd1-7e4c9c63b77a tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 570.259184] env[63088]: DEBUG nova.compute.manager [-] [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 570.259272] env[63088]: DEBUG nova.network.neutron [-] [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 570.452225] env[63088]: DEBUG nova.network.neutron [-] [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 570.493500] env[63088]: DEBUG nova.compute.utils [None req-6e35ce22-cf37-483a-8093-9574de222716 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 570.495091] env[63088]: DEBUG nova.compute.manager [None req-6e35ce22-cf37-483a-8093-9574de222716 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 570.495315] env[63088]: DEBUG nova.network.neutron [None req-6e35ce22-cf37-483a-8093-9574de222716 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 570.569809] env[63088]: INFO nova.compute.manager [-] [instance: 4d5293a0-6945-4731-afb6-996a142c8447] Took 1.05 seconds to deallocate network for instance. [ 570.573564] env[63088]: DEBUG nova.compute.claims [None req-6568dcbe-ebf7-46f4-8cbf-a6cec6f2c264 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] [instance: 4d5293a0-6945-4731-afb6-996a142c8447] Aborting claim: {{(pid=63088) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 570.573564] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6568dcbe-ebf7-46f4-8cbf-a6cec6f2c264 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 570.617865] env[63088]: DEBUG nova.policy [None req-6e35ce22-cf37-483a-8093-9574de222716 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '95f3328a40e34d89ae4b07be1c364cfc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd101ee89dcc143e1876d2217b09cef11', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 570.723644] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1284775, 'name': CreateVM_Task} progress is 99%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 570.849399] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fc8b9b1-1446-49bf-b879-5d1a4b198849 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.854613] env[63088]: DEBUG nova.network.neutron [req-c15b05cf-c77d-446a-a7b3-25e7c22d9a3f req-1111477b-a942-428a-a716-64128cbda1b3 service nova] [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 570.859464] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26c06f02-3d52-4c1f-9906-71b7ac2acb40 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.892321] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13d15fb7-d7a2-4227-b012-717b28c914e1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.900041] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-348e2c47-b20f-4f74-8f05-920cecab0b95 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.913739] env[63088]: DEBUG nova.compute.provider_tree [None req-96b6b7c1-255c-445f-b8db-9f4849d061fb tempest-ImagesNegativeTestJSON-1434405362 tempest-ImagesNegativeTestJSON-1434405362-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 571.003066] env[63088]: DEBUG nova.compute.manager [None req-6e35ce22-cf37-483a-8093-9574de222716 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 571.220514] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1284775, 'name': CreateVM_Task, 'duration_secs': 1.282057} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 571.220719] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c4777721-3f65-455f-9973-c1ed0732de34] Created VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 571.221640] env[63088]: DEBUG oslo_vmware.service [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5c4bf16-40cb-4b0e-8eeb-b500c5efa2bc {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.231158] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 571.235327] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 571.236483] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 571.236483] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e8f9e1f7-a097-4217-b189-00ecf058f62a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.240158] env[63088]: DEBUG nova.network.neutron [req-c15b05cf-c77d-446a-a7b3-25e7c22d9a3f req-1111477b-a942-428a-a716-64128cbda1b3 service nova] [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 571.243035] env[63088]: INFO nova.compute.manager [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Rebuilding instance [ 571.249433] env[63088]: DEBUG oslo_vmware.api [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Waiting for the task: (returnval){ [ 571.249433] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]525e0a8d-61d9-d81b-0750-38bf7893a737" [ 571.249433] env[63088]: _type = "Task" [ 571.249433] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 571.263054] env[63088]: DEBUG oslo_vmware.api [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]525e0a8d-61d9-d81b-0750-38bf7893a737, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 571.297498] env[63088]: DEBUG nova.compute.manager [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 571.298399] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85b30429-ea8c-4091-bc25-36e5a2ae71ca {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.419485] env[63088]: DEBUG nova.scheduler.client.report [None req-96b6b7c1-255c-445f-b8db-9f4849d061fb tempest-ImagesNegativeTestJSON-1434405362 tempest-ImagesNegativeTestJSON-1434405362-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 571.494260] env[63088]: DEBUG nova.network.neutron [None req-6e35ce22-cf37-483a-8093-9574de222716 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] Successfully created port: 27f3a8b8-19ea-48a7-8323-608f6af9414c {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 571.499584] env[63088]: DEBUG nova.compute.manager [req-b259909d-451f-49e7-beac-73421cd9c980 req-72f467cd-3efe-4dc4-8094-e4ae32179895 service nova] [instance: 173fccf1-f258-4f0c-a968-466a27406c55] Received event network-vif-deleted-06dbc948-3bcb-4bb1-9083-7a5a312da444 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 571.499914] env[63088]: DEBUG nova.compute.manager [req-b259909d-451f-49e7-beac-73421cd9c980 req-72f467cd-3efe-4dc4-8094-e4ae32179895 service nova] [instance: 4d5293a0-6945-4731-afb6-996a142c8447] Received event network-changed-0c9febcc-5137-47a2-a0b9-df090f78c2e7 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 571.500174] env[63088]: DEBUG nova.compute.manager [req-b259909d-451f-49e7-beac-73421cd9c980 req-72f467cd-3efe-4dc4-8094-e4ae32179895 service nova] [instance: 4d5293a0-6945-4731-afb6-996a142c8447] Refreshing instance network info cache due to event network-changed-0c9febcc-5137-47a2-a0b9-df090f78c2e7. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 571.500532] env[63088]: DEBUG oslo_concurrency.lockutils [req-b259909d-451f-49e7-beac-73421cd9c980 req-72f467cd-3efe-4dc4-8094-e4ae32179895 service nova] Acquiring lock "refresh_cache-4d5293a0-6945-4731-afb6-996a142c8447" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 571.500532] env[63088]: DEBUG oslo_concurrency.lockutils [req-b259909d-451f-49e7-beac-73421cd9c980 req-72f467cd-3efe-4dc4-8094-e4ae32179895 service nova] Acquired lock "refresh_cache-4d5293a0-6945-4731-afb6-996a142c8447" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 571.500669] env[63088]: DEBUG nova.network.neutron [req-b259909d-451f-49e7-beac-73421cd9c980 req-72f467cd-3efe-4dc4-8094-e4ae32179895 service nova] [instance: 4d5293a0-6945-4731-afb6-996a142c8447] Refreshing network info cache for port 0c9febcc-5137-47a2-a0b9-df090f78c2e7 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 571.748568] env[63088]: DEBUG oslo_concurrency.lockutils [req-c15b05cf-c77d-446a-a7b3-25e7c22d9a3f req-1111477b-a942-428a-a716-64128cbda1b3 service nova] Releasing lock "refresh_cache-5c3f72e3-d4f4-4c34-92a3-075536b85c99" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 571.764656] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 571.764875] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] [instance: c4777721-3f65-455f-9973-c1ed0732de34] Processing image 1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 571.765160] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 571.765304] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 571.765528] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 571.766226] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-193b50fd-bb93-4d50-93c4-18ed873223e9 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.785401] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 571.785648] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 571.786444] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3391b0d-b025-479e-9247-5ff8c91c0bbe {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.794219] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1c4537e6-badf-4911-8cbe-ec39043dd94c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.800524] env[63088]: DEBUG oslo_vmware.api [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Waiting for the task: (returnval){ [ 571.800524] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]527974e8-0c0d-f8db-ff38-c8a34bcd5fb3" [ 571.800524] env[63088]: _type = "Task" [ 571.800524] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 571.808368] env[63088]: DEBUG oslo_vmware.api [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]527974e8-0c0d-f8db-ff38-c8a34bcd5fb3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 571.813180] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 571.813180] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8ba18715-6351-4af9-98f3-48db18c1b895 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.820147] env[63088]: DEBUG oslo_vmware.api [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Waiting for the task: (returnval){ [ 571.820147] env[63088]: value = "task-1284776" [ 571.820147] env[63088]: _type = "Task" [ 571.820147] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 571.826220] env[63088]: DEBUG oslo_vmware.api [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Task: {'id': task-1284776, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 571.928998] env[63088]: DEBUG oslo_concurrency.lockutils [None req-96b6b7c1-255c-445f-b8db-9f4849d061fb tempest-ImagesNegativeTestJSON-1434405362 tempest-ImagesNegativeTestJSON-1434405362-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.938s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 571.928998] env[63088]: ERROR nova.compute.manager [None req-96b6b7c1-255c-445f-b8db-9f4849d061fb tempest-ImagesNegativeTestJSON-1434405362 tempest-ImagesNegativeTestJSON-1434405362-project-member] [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 946407e1-ac24-4d57-9f2f-f2c0994ff966, please check neutron logs for more information. [ 571.928998] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] Traceback (most recent call last): [ 571.928998] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 571.928998] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] self.driver.spawn(context, instance, image_meta, [ 571.928998] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 571.928998] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] self._vmops.spawn(context, instance, image_meta, injected_files, [ 571.928998] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 571.928998] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] vm_ref = self.build_virtual_machine(instance, [ 571.929262] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 571.929262] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] vif_infos = vmwarevif.get_vif_info(self._session, [ 571.929262] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 571.929262] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] for vif in network_info: [ 571.929262] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 571.929262] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] return self._sync_wrapper(fn, *args, **kwargs) [ 571.929262] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 571.929262] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] self.wait() [ 571.929262] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 571.929262] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] self[:] = self._gt.wait() [ 571.929262] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 571.929262] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] return self._exit_event.wait() [ 571.929262] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 571.929581] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] result = hub.switch() [ 571.929581] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 571.929581] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] return self.greenlet.switch() [ 571.929581] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 571.929581] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] result = function(*args, **kwargs) [ 571.929581] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 571.929581] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] return func(*args, **kwargs) [ 571.929581] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 571.929581] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] raise e [ 571.929581] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 571.929581] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] nwinfo = self.network_api.allocate_for_instance( [ 571.929581] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 571.929581] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] created_port_ids = self._update_ports_for_instance( [ 571.929894] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 571.929894] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] with excutils.save_and_reraise_exception(): [ 571.929894] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 571.929894] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] self.force_reraise() [ 571.929894] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 571.929894] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] raise self.value [ 571.929894] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 571.929894] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] updated_port = self._update_port( [ 571.929894] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 571.929894] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] _ensure_no_port_binding_failure(port) [ 571.929894] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 571.929894] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] raise exception.PortBindingFailed(port_id=port['id']) [ 571.930193] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] nova.exception.PortBindingFailed: Binding failed for port 946407e1-ac24-4d57-9f2f-f2c0994ff966, please check neutron logs for more information. [ 571.930193] env[63088]: ERROR nova.compute.manager [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] [ 571.930193] env[63088]: DEBUG nova.compute.utils [None req-96b6b7c1-255c-445f-b8db-9f4849d061fb tempest-ImagesNegativeTestJSON-1434405362 tempest-ImagesNegativeTestJSON-1434405362-project-member] [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] Binding failed for port 946407e1-ac24-4d57-9f2f-f2c0994ff966, please check neutron logs for more information. {{(pid=63088) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 571.931910] env[63088]: DEBUG oslo_concurrency.lockutils [None req-fa54bfad-a3c9-4ca3-8f38-8bc8499d1488 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.661s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 571.943929] env[63088]: DEBUG nova.compute.manager [None req-96b6b7c1-255c-445f-b8db-9f4849d061fb tempest-ImagesNegativeTestJSON-1434405362 tempest-ImagesNegativeTestJSON-1434405362-project-member] [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] Build of instance 0fd65b6b-2363-4b45-bc07-365157a8b052 was re-scheduled: Binding failed for port 946407e1-ac24-4d57-9f2f-f2c0994ff966, please check neutron logs for more information. {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 571.943929] env[63088]: DEBUG nova.compute.manager [None req-96b6b7c1-255c-445f-b8db-9f4849d061fb tempest-ImagesNegativeTestJSON-1434405362 tempest-ImagesNegativeTestJSON-1434405362-project-member] [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] Unplugging VIFs for instance {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 571.943929] env[63088]: DEBUG oslo_concurrency.lockutils [None req-96b6b7c1-255c-445f-b8db-9f4849d061fb tempest-ImagesNegativeTestJSON-1434405362 tempest-ImagesNegativeTestJSON-1434405362-project-member] Acquiring lock "refresh_cache-0fd65b6b-2363-4b45-bc07-365157a8b052" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 571.943929] env[63088]: DEBUG oslo_concurrency.lockutils [None req-96b6b7c1-255c-445f-b8db-9f4849d061fb tempest-ImagesNegativeTestJSON-1434405362 tempest-ImagesNegativeTestJSON-1434405362-project-member] Acquired lock "refresh_cache-0fd65b6b-2363-4b45-bc07-365157a8b052" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 571.944212] env[63088]: DEBUG nova.network.neutron [None req-96b6b7c1-255c-445f-b8db-9f4849d061fb tempest-ImagesNegativeTestJSON-1434405362 tempest-ImagesNegativeTestJSON-1434405362-project-member] [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 572.014213] env[63088]: DEBUG nova.compute.manager [None req-6e35ce22-cf37-483a-8093-9574de222716 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 572.050801] env[63088]: DEBUG nova.network.neutron [req-b259909d-451f-49e7-beac-73421cd9c980 req-72f467cd-3efe-4dc4-8094-e4ae32179895 service nova] [instance: 4d5293a0-6945-4731-afb6-996a142c8447] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 572.054417] env[63088]: DEBUG nova.virt.hardware [None req-6e35ce22-cf37-483a-8093-9574de222716 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 572.055477] env[63088]: DEBUG nova.virt.hardware [None req-6e35ce22-cf37-483a-8093-9574de222716 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 572.057264] env[63088]: DEBUG nova.virt.hardware [None req-6e35ce22-cf37-483a-8093-9574de222716 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 572.057264] env[63088]: DEBUG nova.virt.hardware [None req-6e35ce22-cf37-483a-8093-9574de222716 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 572.057264] env[63088]: DEBUG nova.virt.hardware [None req-6e35ce22-cf37-483a-8093-9574de222716 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 572.057264] env[63088]: DEBUG nova.virt.hardware [None req-6e35ce22-cf37-483a-8093-9574de222716 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 572.057264] env[63088]: DEBUG nova.virt.hardware [None req-6e35ce22-cf37-483a-8093-9574de222716 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 572.057635] env[63088]: DEBUG nova.virt.hardware [None req-6e35ce22-cf37-483a-8093-9574de222716 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 572.057635] env[63088]: DEBUG nova.virt.hardware [None req-6e35ce22-cf37-483a-8093-9574de222716 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 572.057635] env[63088]: DEBUG nova.virt.hardware [None req-6e35ce22-cf37-483a-8093-9574de222716 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 572.057635] env[63088]: DEBUG nova.virt.hardware [None req-6e35ce22-cf37-483a-8093-9574de222716 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 572.057937] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52ab6ec8-4820-44d8-ae31-11a7478e319e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.067220] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-632b8b5c-16e6-490f-ad89-7144bc2f1843 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.315259] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] [instance: c4777721-3f65-455f-9973-c1ed0732de34] Preparing fetch location {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 572.315530] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Creating directory with path [datastore2] vmware_temp/ab951fe9-c5dc-4a8c-9563-9985e88c030e/1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 572.315772] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-21c68029-6854-42fd-b266-1938dcd147f8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.320465] env[63088]: DEBUG nova.network.neutron [req-b259909d-451f-49e7-beac-73421cd9c980 req-72f467cd-3efe-4dc4-8094-e4ae32179895 service nova] [instance: 4d5293a0-6945-4731-afb6-996a142c8447] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 572.334905] env[63088]: DEBUG oslo_vmware.api [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Task: {'id': task-1284776, 'name': PowerOffVM_Task, 'duration_secs': 0.122025} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 572.335195] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 572.335412] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 572.337210] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8865a3f-149b-448d-afbe-233473d5182a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.342081] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Created directory with path [datastore2] vmware_temp/ab951fe9-c5dc-4a8c-9563-9985e88c030e/1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 572.342266] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] [instance: c4777721-3f65-455f-9973-c1ed0732de34] Fetch image to [datastore2] vmware_temp/ab951fe9-c5dc-4a8c-9563-9985e88c030e/1e8c5d18-0a03-4e18-afe1-de5a6e255953/tmp-sparse.vmdk {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 572.342434] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] [instance: c4777721-3f65-455f-9973-c1ed0732de34] Downloading image file data 1e8c5d18-0a03-4e18-afe1-de5a6e255953 to [datastore2] vmware_temp/ab951fe9-c5dc-4a8c-9563-9985e88c030e/1e8c5d18-0a03-4e18-afe1-de5a6e255953/tmp-sparse.vmdk on the data store datastore2 {{(pid=63088) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 572.350620] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-896d1b42-9b6a-4948-949e-0c011a3edacf {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.353576] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Unregistering the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 572.353812] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4152f72d-0f39-48bb-a145-d4464660e6b6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.359737] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da7c67ac-cdd7-4cf5-bdbc-88528fa8577a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.373348] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5993810-d9c8-4bb9-9d5a-f71581e7bf15 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.412821] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c622298e-2983-4b19-98f3-94214f170178 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.415126] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Unregistered the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 572.415348] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Deleting contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 572.415653] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Deleting the datastore file [datastore1] e8cfbce4-cfb2-463e-9f38-78a058dab6e1 {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 572.415912] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-66f8cdad-f49d-4c67-8a14-4a1758c40b54 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.421715] env[63088]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-c1e7d091-2255-4366-9106-7cae4d2e3d3f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.424947] env[63088]: DEBUG oslo_vmware.api [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Waiting for the task: (returnval){ [ 572.424947] env[63088]: value = "task-1284778" [ 572.424947] env[63088]: _type = "Task" [ 572.424947] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 572.433226] env[63088]: DEBUG oslo_vmware.api [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Task: {'id': task-1284778, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 572.440280] env[63088]: DEBUG nova.network.neutron [-] [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 572.478831] env[63088]: DEBUG nova.network.neutron [None req-96b6b7c1-255c-445f-b8db-9f4849d061fb tempest-ImagesNegativeTestJSON-1434405362 tempest-ImagesNegativeTestJSON-1434405362-project-member] [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 572.516298] env[63088]: DEBUG nova.virt.vmwareapi.images [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] [instance: c4777721-3f65-455f-9973-c1ed0732de34] Downloading image file data 1e8c5d18-0a03-4e18-afe1-de5a6e255953 to the data store datastore2 {{(pid=63088) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 572.588300] env[63088]: DEBUG oslo_vmware.rw_handles [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/ab951fe9-c5dc-4a8c-9563-9985e88c030e/1e8c5d18-0a03-4e18-afe1-de5a6e255953/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=63088) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 572.787411] env[63088]: DEBUG nova.network.neutron [None req-96b6b7c1-255c-445f-b8db-9f4849d061fb tempest-ImagesNegativeTestJSON-1434405362 tempest-ImagesNegativeTestJSON-1434405362-project-member] [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 572.828209] env[63088]: DEBUG oslo_concurrency.lockutils [req-b259909d-451f-49e7-beac-73421cd9c980 req-72f467cd-3efe-4dc4-8094-e4ae32179895 service nova] Releasing lock "refresh_cache-4d5293a0-6945-4731-afb6-996a142c8447" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 572.829032] env[63088]: DEBUG nova.compute.manager [req-b259909d-451f-49e7-beac-73421cd9c980 req-72f467cd-3efe-4dc4-8094-e4ae32179895 service nova] [instance: 4d5293a0-6945-4731-afb6-996a142c8447] Received event network-vif-deleted-0c9febcc-5137-47a2-a0b9-df090f78c2e7 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 572.838385] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbcfa941-297a-472c-835f-40fd73543b35 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.854016] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61e653cf-7a57-4b1e-bbeb-56ac08a0fdc6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.893135] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-863c1026-070d-4601-9c89-d3a3385f1fc4 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.907018] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2ecd0a4-1772-44e8-b53b-dca7869bd5f9 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.922875] env[63088]: DEBUG nova.compute.provider_tree [None req-fa54bfad-a3c9-4ca3-8f38-8bc8499d1488 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 572.930386] env[63088]: DEBUG nova.compute.manager [req-0e62c0a2-924e-4b6d-9031-16516437eb40 req-d9bc1053-c468-4847-93b2-541fdc12569a service nova] [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] Received event network-vif-deleted-010a5358-1268-49c2-b2b4-9525552d6fdc {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 572.940263] env[63088]: DEBUG oslo_vmware.api [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Task: {'id': task-1284778, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.148375} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 572.940499] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 572.940677] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Deleted contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 572.940844] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 572.943242] env[63088]: INFO nova.compute.manager [-] [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] Took 2.68 seconds to deallocate network for instance. [ 572.948641] env[63088]: DEBUG nova.compute.claims [None req-bc610745-9520-42a6-9fd1-7e4c9c63b77a tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] Aborting claim: {{(pid=63088) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 572.950184] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bc610745-9520-42a6-9fd1-7e4c9c63b77a tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 573.261964] env[63088]: DEBUG oslo_vmware.rw_handles [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Completed reading data from the image iterator. {{(pid=63088) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 573.262342] env[63088]: DEBUG oslo_vmware.rw_handles [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Closing write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/ab951fe9-c5dc-4a8c-9563-9985e88c030e/1e8c5d18-0a03-4e18-afe1-de5a6e255953/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=63088) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 573.291610] env[63088]: DEBUG oslo_concurrency.lockutils [None req-96b6b7c1-255c-445f-b8db-9f4849d061fb tempest-ImagesNegativeTestJSON-1434405362 tempest-ImagesNegativeTestJSON-1434405362-project-member] Releasing lock "refresh_cache-0fd65b6b-2363-4b45-bc07-365157a8b052" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 573.291764] env[63088]: DEBUG nova.compute.manager [None req-96b6b7c1-255c-445f-b8db-9f4849d061fb tempest-ImagesNegativeTestJSON-1434405362 tempest-ImagesNegativeTestJSON-1434405362-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 573.291875] env[63088]: DEBUG nova.compute.manager [None req-96b6b7c1-255c-445f-b8db-9f4849d061fb tempest-ImagesNegativeTestJSON-1434405362 tempest-ImagesNegativeTestJSON-1434405362-project-member] [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 573.292089] env[63088]: DEBUG nova.network.neutron [None req-96b6b7c1-255c-445f-b8db-9f4849d061fb tempest-ImagesNegativeTestJSON-1434405362 tempest-ImagesNegativeTestJSON-1434405362-project-member] [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 573.312163] env[63088]: DEBUG nova.network.neutron [None req-96b6b7c1-255c-445f-b8db-9f4849d061fb tempest-ImagesNegativeTestJSON-1434405362 tempest-ImagesNegativeTestJSON-1434405362-project-member] [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 573.395539] env[63088]: ERROR nova.compute.manager [None req-6e35ce22-cf37-483a-8093-9574de222716 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 27f3a8b8-19ea-48a7-8323-608f6af9414c, please check neutron logs for more information. [ 573.395539] env[63088]: ERROR nova.compute.manager Traceback (most recent call last): [ 573.395539] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 573.395539] env[63088]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 573.395539] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 573.395539] env[63088]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 573.395539] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 573.395539] env[63088]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 573.395539] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 573.395539] env[63088]: ERROR nova.compute.manager self.force_reraise() [ 573.395539] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 573.395539] env[63088]: ERROR nova.compute.manager raise self.value [ 573.395539] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 573.395539] env[63088]: ERROR nova.compute.manager updated_port = self._update_port( [ 573.395539] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 573.395539] env[63088]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 573.396380] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 573.396380] env[63088]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 573.396380] env[63088]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 27f3a8b8-19ea-48a7-8323-608f6af9414c, please check neutron logs for more information. [ 573.396380] env[63088]: ERROR nova.compute.manager [ 573.396380] env[63088]: Traceback (most recent call last): [ 573.396380] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 573.396380] env[63088]: listener.cb(fileno) [ 573.396380] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 573.396380] env[63088]: result = function(*args, **kwargs) [ 573.396380] env[63088]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 573.396380] env[63088]: return func(*args, **kwargs) [ 573.396380] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 573.396380] env[63088]: raise e [ 573.396380] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 573.396380] env[63088]: nwinfo = self.network_api.allocate_for_instance( [ 573.396380] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 573.396380] env[63088]: created_port_ids = self._update_ports_for_instance( [ 573.396380] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 573.396380] env[63088]: with excutils.save_and_reraise_exception(): [ 573.396380] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 573.396380] env[63088]: self.force_reraise() [ 573.396380] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 573.396380] env[63088]: raise self.value [ 573.396380] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 573.396380] env[63088]: updated_port = self._update_port( [ 573.396380] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 573.396380] env[63088]: _ensure_no_port_binding_failure(port) [ 573.396380] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 573.396380] env[63088]: raise exception.PortBindingFailed(port_id=port['id']) [ 573.397201] env[63088]: nova.exception.PortBindingFailed: Binding failed for port 27f3a8b8-19ea-48a7-8323-608f6af9414c, please check neutron logs for more information. [ 573.397201] env[63088]: Removing descriptor: 17 [ 573.397201] env[63088]: ERROR nova.compute.manager [None req-6e35ce22-cf37-483a-8093-9574de222716 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 27f3a8b8-19ea-48a7-8323-608f6af9414c, please check neutron logs for more information. [ 573.397201] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] Traceback (most recent call last): [ 573.397201] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 573.397201] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] yield resources [ 573.397201] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 573.397201] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] self.driver.spawn(context, instance, image_meta, [ 573.397201] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 573.397201] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] self._vmops.spawn(context, instance, image_meta, injected_files, [ 573.397201] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 573.397201] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] vm_ref = self.build_virtual_machine(instance, [ 573.397621] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 573.397621] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] vif_infos = vmwarevif.get_vif_info(self._session, [ 573.397621] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 573.397621] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] for vif in network_info: [ 573.397621] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 573.397621] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] return self._sync_wrapper(fn, *args, **kwargs) [ 573.397621] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 573.397621] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] self.wait() [ 573.397621] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 573.397621] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] self[:] = self._gt.wait() [ 573.397621] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 573.397621] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] return self._exit_event.wait() [ 573.397621] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 573.397996] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] result = hub.switch() [ 573.397996] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 573.397996] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] return self.greenlet.switch() [ 573.397996] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 573.397996] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] result = function(*args, **kwargs) [ 573.397996] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 573.397996] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] return func(*args, **kwargs) [ 573.397996] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 573.397996] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] raise e [ 573.397996] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 573.397996] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] nwinfo = self.network_api.allocate_for_instance( [ 573.397996] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 573.397996] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] created_port_ids = self._update_ports_for_instance( [ 573.398342] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 573.398342] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] with excutils.save_and_reraise_exception(): [ 573.398342] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 573.398342] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] self.force_reraise() [ 573.398342] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 573.398342] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] raise self.value [ 573.398342] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 573.398342] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] updated_port = self._update_port( [ 573.398342] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 573.398342] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] _ensure_no_port_binding_failure(port) [ 573.398342] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 573.398342] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] raise exception.PortBindingFailed(port_id=port['id']) [ 573.398675] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] nova.exception.PortBindingFailed: Binding failed for port 27f3a8b8-19ea-48a7-8323-608f6af9414c, please check neutron logs for more information. [ 573.398675] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] [ 573.398675] env[63088]: INFO nova.compute.manager [None req-6e35ce22-cf37-483a-8093-9574de222716 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] Terminating instance [ 573.398675] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6e35ce22-cf37-483a-8093-9574de222716 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] Acquiring lock "refresh_cache-44e39b2b-2bac-4a6e-9534-d2774d094931" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 573.398675] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6e35ce22-cf37-483a-8093-9574de222716 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] Acquired lock "refresh_cache-44e39b2b-2bac-4a6e-9534-d2774d094931" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 573.398675] env[63088]: DEBUG nova.network.neutron [None req-6e35ce22-cf37-483a-8093-9574de222716 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 573.408508] env[63088]: DEBUG nova.virt.vmwareapi.images [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] [instance: c4777721-3f65-455f-9973-c1ed0732de34] Downloaded image file data 1e8c5d18-0a03-4e18-afe1-de5a6e255953 to vmware_temp/ab951fe9-c5dc-4a8c-9563-9985e88c030e/1e8c5d18-0a03-4e18-afe1-de5a6e255953/tmp-sparse.vmdk on the data store datastore2 {{(pid=63088) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 573.410335] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] [instance: c4777721-3f65-455f-9973-c1ed0732de34] Caching image {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 573.410529] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Copying Virtual Disk [datastore2] vmware_temp/ab951fe9-c5dc-4a8c-9563-9985e88c030e/1e8c5d18-0a03-4e18-afe1-de5a6e255953/tmp-sparse.vmdk to [datastore2] vmware_temp/ab951fe9-c5dc-4a8c-9563-9985e88c030e/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 573.411128] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-075719f9-a874-47d3-b97d-ceec6991a6b8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.418759] env[63088]: DEBUG oslo_vmware.api [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Waiting for the task: (returnval){ [ 573.418759] env[63088]: value = "task-1284779" [ 573.418759] env[63088]: _type = "Task" [ 573.418759] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 573.428963] env[63088]: DEBUG oslo_vmware.api [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Task: {'id': task-1284779, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 573.432178] env[63088]: DEBUG nova.scheduler.client.report [None req-fa54bfad-a3c9-4ca3-8f38-8bc8499d1488 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 573.814197] env[63088]: DEBUG nova.network.neutron [None req-96b6b7c1-255c-445f-b8db-9f4849d061fb tempest-ImagesNegativeTestJSON-1434405362 tempest-ImagesNegativeTestJSON-1434405362-project-member] [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 573.918686] env[63088]: DEBUG nova.network.neutron [None req-6e35ce22-cf37-483a-8093-9574de222716 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 573.935046] env[63088]: DEBUG oslo_vmware.api [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Task: {'id': task-1284779, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 573.940715] env[63088]: DEBUG oslo_concurrency.lockutils [None req-fa54bfad-a3c9-4ca3-8f38-8bc8499d1488 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.013s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 573.941398] env[63088]: ERROR nova.compute.manager [None req-fa54bfad-a3c9-4ca3-8f38-8bc8499d1488 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f66bce24-6dbd-4128-9ab8-86cf1c74d872, please check neutron logs for more information. [ 573.941398] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] Traceback (most recent call last): [ 573.941398] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 573.941398] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] self.driver.spawn(context, instance, image_meta, [ 573.941398] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 573.941398] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 573.941398] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 573.941398] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] vm_ref = self.build_virtual_machine(instance, [ 573.941398] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 573.941398] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] vif_infos = vmwarevif.get_vif_info(self._session, [ 573.941398] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 573.941762] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] for vif in network_info: [ 573.941762] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 573.941762] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] return self._sync_wrapper(fn, *args, **kwargs) [ 573.941762] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 573.941762] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] self.wait() [ 573.941762] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 573.941762] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] self[:] = self._gt.wait() [ 573.941762] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 573.941762] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] return self._exit_event.wait() [ 573.941762] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 573.941762] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] result = hub.switch() [ 573.941762] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 573.941762] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] return self.greenlet.switch() [ 573.942167] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 573.942167] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] result = function(*args, **kwargs) [ 573.942167] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 573.942167] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] return func(*args, **kwargs) [ 573.942167] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 573.942167] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] raise e [ 573.942167] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 573.942167] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] nwinfo = self.network_api.allocate_for_instance( [ 573.942167] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 573.942167] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] created_port_ids = self._update_ports_for_instance( [ 573.942167] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 573.942167] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] with excutils.save_and_reraise_exception(): [ 573.942167] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 573.942515] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] self.force_reraise() [ 573.942515] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 573.942515] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] raise self.value [ 573.942515] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 573.942515] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] updated_port = self._update_port( [ 573.942515] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 573.942515] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] _ensure_no_port_binding_failure(port) [ 573.942515] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 573.942515] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] raise exception.PortBindingFailed(port_id=port['id']) [ 573.942515] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] nova.exception.PortBindingFailed: Binding failed for port f66bce24-6dbd-4128-9ab8-86cf1c74d872, please check neutron logs for more information. [ 573.942515] env[63088]: ERROR nova.compute.manager [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] [ 573.943060] env[63088]: DEBUG nova.compute.utils [None req-fa54bfad-a3c9-4ca3-8f38-8bc8499d1488 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] Binding failed for port f66bce24-6dbd-4128-9ab8-86cf1c74d872, please check neutron logs for more information. {{(pid=63088) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 573.943653] env[63088]: DEBUG oslo_concurrency.lockutils [None req-710306d6-422f-4cdd-91a1-33da212e5f03 tempest-ServersAdminNegativeTestJSON-229078574 tempest-ServersAdminNegativeTestJSON-229078574-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.718s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 573.946532] env[63088]: DEBUG nova.compute.manager [None req-fa54bfad-a3c9-4ca3-8f38-8bc8499d1488 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] Build of instance 05d2ad93-949e-42e3-85e9-eed9baaebe8c was re-scheduled: Binding failed for port f66bce24-6dbd-4128-9ab8-86cf1c74d872, please check neutron logs for more information. {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 573.947107] env[63088]: DEBUG nova.compute.manager [None req-fa54bfad-a3c9-4ca3-8f38-8bc8499d1488 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] Unplugging VIFs for instance {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 573.947374] env[63088]: DEBUG oslo_concurrency.lockutils [None req-fa54bfad-a3c9-4ca3-8f38-8bc8499d1488 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] Acquiring lock "refresh_cache-05d2ad93-949e-42e3-85e9-eed9baaebe8c" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 573.947531] env[63088]: DEBUG oslo_concurrency.lockutils [None req-fa54bfad-a3c9-4ca3-8f38-8bc8499d1488 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] Acquired lock "refresh_cache-05d2ad93-949e-42e3-85e9-eed9baaebe8c" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 573.948385] env[63088]: DEBUG nova.network.neutron [None req-fa54bfad-a3c9-4ca3-8f38-8bc8499d1488 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 573.996635] env[63088]: DEBUG nova.virt.hardware [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 573.996905] env[63088]: DEBUG nova.virt.hardware [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 573.997157] env[63088]: DEBUG nova.virt.hardware [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 573.997910] env[63088]: DEBUG nova.virt.hardware [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 573.997910] env[63088]: DEBUG nova.virt.hardware [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 573.997910] env[63088]: DEBUG nova.virt.hardware [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 573.997910] env[63088]: DEBUG nova.virt.hardware [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 573.998232] env[63088]: DEBUG nova.virt.hardware [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 573.998428] env[63088]: DEBUG nova.virt.hardware [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 573.998488] env[63088]: DEBUG nova.virt.hardware [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 573.998663] env[63088]: DEBUG nova.virt.hardware [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 573.999885] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1817bcca-c7fd-4cf7-966e-f9d14de4f5d7 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.008623] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d32375e3-2b7b-4c96-8df7-e481ef842848 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.028292] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Instance VIF info [] {{(pid=63088) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 574.036566] env[63088]: DEBUG oslo.service.loopingcall [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 574.038199] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Creating VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 574.038199] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-232a436d-5a44-401a-a2f4-c40d09b73472 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.058918] env[63088]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 574.058918] env[63088]: value = "task-1284780" [ 574.058918] env[63088]: _type = "Task" [ 574.058918] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 574.070519] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1284780, 'name': CreateVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 574.084309] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f5f50761-0430-407c-84cc-7b2c9cf8683b tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] Acquiring lock "f29f6a5e-8621-4802-a679-d5c9a8fb2461" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 574.084309] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f5f50761-0430-407c-84cc-7b2c9cf8683b tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] Lock "f29f6a5e-8621-4802-a679-d5c9a8fb2461" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 574.085599] env[63088]: DEBUG nova.network.neutron [None req-6e35ce22-cf37-483a-8093-9574de222716 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 574.317954] env[63088]: INFO nova.compute.manager [None req-96b6b7c1-255c-445f-b8db-9f4849d061fb tempest-ImagesNegativeTestJSON-1434405362 tempest-ImagesNegativeTestJSON-1434405362-project-member] [instance: 0fd65b6b-2363-4b45-bc07-365157a8b052] Took 1.03 seconds to deallocate network for instance. [ 574.436119] env[63088]: DEBUG oslo_vmware.api [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Task: {'id': task-1284779, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.699382} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 574.436508] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Copied Virtual Disk [datastore2] vmware_temp/ab951fe9-c5dc-4a8c-9563-9985e88c030e/1e8c5d18-0a03-4e18-afe1-de5a6e255953/tmp-sparse.vmdk to [datastore2] vmware_temp/ab951fe9-c5dc-4a8c-9563-9985e88c030e/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 574.436719] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Deleting the datastore file [datastore2] vmware_temp/ab951fe9-c5dc-4a8c-9563-9985e88c030e/1e8c5d18-0a03-4e18-afe1-de5a6e255953/tmp-sparse.vmdk {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 574.437010] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ac446c82-3319-4f2f-8a0c-588231b0016a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.449982] env[63088]: DEBUG oslo_vmware.api [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Waiting for the task: (returnval){ [ 574.449982] env[63088]: value = "task-1284781" [ 574.449982] env[63088]: _type = "Task" [ 574.449982] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 574.465943] env[63088]: DEBUG oslo_vmware.api [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Task: {'id': task-1284781, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 574.486269] env[63088]: DEBUG nova.network.neutron [None req-fa54bfad-a3c9-4ca3-8f38-8bc8499d1488 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 574.555163] env[63088]: DEBUG nova.network.neutron [None req-fa54bfad-a3c9-4ca3-8f38-8bc8499d1488 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 574.573334] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1284780, 'name': CreateVM_Task, 'duration_secs': 0.308141} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 574.576317] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Created VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 574.577874] env[63088]: DEBUG oslo_concurrency.lockutils [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 574.578119] env[63088]: DEBUG oslo_concurrency.lockutils [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 574.578533] env[63088]: DEBUG oslo_concurrency.lockutils [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 574.579026] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7953639b-0f83-41a6-a04c-86dcec79fb5a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.585831] env[63088]: DEBUG oslo_vmware.api [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Waiting for the task: (returnval){ [ 574.585831] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]522277e0-70bf-6adb-b795-667079bc0cc2" [ 574.585831] env[63088]: _type = "Task" [ 574.585831] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 574.590812] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6e35ce22-cf37-483a-8093-9574de222716 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] Releasing lock "refresh_cache-44e39b2b-2bac-4a6e-9534-d2774d094931" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 574.596636] env[63088]: DEBUG nova.compute.manager [None req-6e35ce22-cf37-483a-8093-9574de222716 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 574.596636] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-6e35ce22-cf37-483a-8093-9574de222716 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 574.596990] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5d863dd7-9e86-43ea-854c-f54df3347e75 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.610107] env[63088]: DEBUG oslo_concurrency.lockutils [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 574.610107] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Processing image 1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 574.610107] env[63088]: DEBUG oslo_concurrency.lockutils [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 574.612271] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b7e3188-bd2a-4ace-a91b-5cd56e68fdee {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.633107] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b1b885bf-cef9-4808-8f16-707ab04499b3 tempest-ServersTestManualDisk-1859582357 tempest-ServersTestManualDisk-1859582357-project-member] Acquiring lock "498c8e1e-18c6-4e09-a90d-106600bed2cd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 574.633107] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b1b885bf-cef9-4808-8f16-707ab04499b3 tempest-ServersTestManualDisk-1859582357 tempest-ServersTestManualDisk-1859582357-project-member] Lock "498c8e1e-18c6-4e09-a90d-106600bed2cd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 574.645642] env[63088]: WARNING nova.virt.vmwareapi.vmops [None req-6e35ce22-cf37-483a-8093-9574de222716 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 44e39b2b-2bac-4a6e-9534-d2774d094931 could not be found. [ 574.645642] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-6e35ce22-cf37-483a-8093-9574de222716 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 574.645865] env[63088]: INFO nova.compute.manager [None req-6e35ce22-cf37-483a-8093-9574de222716 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] Took 0.05 seconds to destroy the instance on the hypervisor. [ 574.646054] env[63088]: DEBUG oslo.service.loopingcall [None req-6e35ce22-cf37-483a-8093-9574de222716 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 574.648816] env[63088]: DEBUG nova.compute.manager [-] [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 574.648948] env[63088]: DEBUG nova.network.neutron [-] [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 574.676536] env[63088]: DEBUG nova.network.neutron [-] [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 574.838158] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f9d3426-7fc3-4023-af22-2296c86544a0 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.847333] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8ff7b8a-d1b9-4bd3-a024-aa25198a257d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.883276] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8994158f-8720-411b-b37b-e13717d772f5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.892416] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f14c78b-d20f-4427-ad59-be5e3661a206 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.908281] env[63088]: DEBUG nova.compute.provider_tree [None req-710306d6-422f-4cdd-91a1-33da212e5f03 tempest-ServersAdminNegativeTestJSON-229078574 tempest-ServersAdminNegativeTestJSON-229078574-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 574.962907] env[63088]: DEBUG oslo_vmware.api [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Task: {'id': task-1284781, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.023539} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 574.963369] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 574.964837] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Moving file from [datastore2] vmware_temp/ab951fe9-c5dc-4a8c-9563-9985e88c030e/1e8c5d18-0a03-4e18-afe1-de5a6e255953 to [datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953. {{(pid=63088) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 574.964837] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-4107ea59-642f-4dc6-9224-9f3ee80c8157 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.972718] env[63088]: DEBUG oslo_vmware.api [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Waiting for the task: (returnval){ [ 574.972718] env[63088]: value = "task-1284782" [ 574.972718] env[63088]: _type = "Task" [ 574.972718] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 574.981854] env[63088]: DEBUG oslo_vmware.api [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Task: {'id': task-1284782, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 575.058341] env[63088]: DEBUG oslo_concurrency.lockutils [None req-fa54bfad-a3c9-4ca3-8f38-8bc8499d1488 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] Releasing lock "refresh_cache-05d2ad93-949e-42e3-85e9-eed9baaebe8c" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 575.058591] env[63088]: DEBUG nova.compute.manager [None req-fa54bfad-a3c9-4ca3-8f38-8bc8499d1488 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 575.058758] env[63088]: DEBUG nova.compute.manager [None req-fa54bfad-a3c9-4ca3-8f38-8bc8499d1488 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 575.058918] env[63088]: DEBUG nova.network.neutron [None req-fa54bfad-a3c9-4ca3-8f38-8bc8499d1488 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 575.080072] env[63088]: DEBUG nova.network.neutron [None req-fa54bfad-a3c9-4ca3-8f38-8bc8499d1488 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 575.179041] env[63088]: DEBUG nova.network.neutron [-] [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 575.349210] env[63088]: INFO nova.scheduler.client.report [None req-96b6b7c1-255c-445f-b8db-9f4849d061fb tempest-ImagesNegativeTestJSON-1434405362 tempest-ImagesNegativeTestJSON-1434405362-project-member] Deleted allocations for instance 0fd65b6b-2363-4b45-bc07-365157a8b052 [ 575.412767] env[63088]: DEBUG nova.scheduler.client.report [None req-710306d6-422f-4cdd-91a1-33da212e5f03 tempest-ServersAdminNegativeTestJSON-229078574 tempest-ServersAdminNegativeTestJSON-229078574-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 575.482303] env[63088]: DEBUG oslo_vmware.api [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Task: {'id': task-1284782, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.025021} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 575.482627] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] File moved {{(pid=63088) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 575.482744] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] [instance: c4777721-3f65-455f-9973-c1ed0732de34] Cleaning up location [datastore2] vmware_temp/ab951fe9-c5dc-4a8c-9563-9985e88c030e {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 575.482906] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Deleting the datastore file [datastore2] vmware_temp/ab951fe9-c5dc-4a8c-9563-9985e88c030e {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 575.483179] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3c946cf3-0c0b-4a04-beb5-118eea7e2b5e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.489997] env[63088]: DEBUG oslo_vmware.api [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Waiting for the task: (returnval){ [ 575.489997] env[63088]: value = "task-1284783" [ 575.489997] env[63088]: _type = "Task" [ 575.489997] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 575.497896] env[63088]: DEBUG oslo_vmware.api [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Task: {'id': task-1284783, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 575.583308] env[63088]: DEBUG nova.network.neutron [None req-fa54bfad-a3c9-4ca3-8f38-8bc8499d1488 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 575.641048] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0564aade-729d-4cf1-a59d-513384386c54 tempest-AttachInterfacesV270Test-1783457293 tempest-AttachInterfacesV270Test-1783457293-project-member] Acquiring lock "421e0f76-024d-4793-8f00-98731bb6b1af" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 575.641709] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0564aade-729d-4cf1-a59d-513384386c54 tempest-AttachInterfacesV270Test-1783457293 tempest-AttachInterfacesV270Test-1783457293-project-member] Lock "421e0f76-024d-4793-8f00-98731bb6b1af" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 575.673128] env[63088]: DEBUG nova.compute.manager [req-cf482d53-8aca-4cbb-88da-85a2ba03e975 req-1439ab49-7837-4c3f-a8a5-1f897672cb51 service nova] [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] Received event network-changed-27f3a8b8-19ea-48a7-8323-608f6af9414c {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 575.673350] env[63088]: DEBUG nova.compute.manager [req-cf482d53-8aca-4cbb-88da-85a2ba03e975 req-1439ab49-7837-4c3f-a8a5-1f897672cb51 service nova] [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] Refreshing instance network info cache due to event network-changed-27f3a8b8-19ea-48a7-8323-608f6af9414c. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 575.673527] env[63088]: DEBUG oslo_concurrency.lockutils [req-cf482d53-8aca-4cbb-88da-85a2ba03e975 req-1439ab49-7837-4c3f-a8a5-1f897672cb51 service nova] Acquiring lock "refresh_cache-44e39b2b-2bac-4a6e-9534-d2774d094931" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 575.673670] env[63088]: DEBUG oslo_concurrency.lockutils [req-cf482d53-8aca-4cbb-88da-85a2ba03e975 req-1439ab49-7837-4c3f-a8a5-1f897672cb51 service nova] Acquired lock "refresh_cache-44e39b2b-2bac-4a6e-9534-d2774d094931" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 575.674066] env[63088]: DEBUG nova.network.neutron [req-cf482d53-8aca-4cbb-88da-85a2ba03e975 req-1439ab49-7837-4c3f-a8a5-1f897672cb51 service nova] [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] Refreshing network info cache for port 27f3a8b8-19ea-48a7-8323-608f6af9414c {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 575.682079] env[63088]: INFO nova.compute.manager [-] [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] Took 1.03 seconds to deallocate network for instance. [ 575.685203] env[63088]: DEBUG nova.compute.claims [None req-6e35ce22-cf37-483a-8093-9574de222716 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] Aborting claim: {{(pid=63088) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 575.685292] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6e35ce22-cf37-483a-8093-9574de222716 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 575.866416] env[63088]: DEBUG oslo_concurrency.lockutils [None req-96b6b7c1-255c-445f-b8db-9f4849d061fb tempest-ImagesNegativeTestJSON-1434405362 tempest-ImagesNegativeTestJSON-1434405362-project-member] Lock "0fd65b6b-2363-4b45-bc07-365157a8b052" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 32.082s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 575.917854] env[63088]: DEBUG oslo_concurrency.lockutils [None req-710306d6-422f-4cdd-91a1-33da212e5f03 tempest-ServersAdminNegativeTestJSON-229078574 tempest-ServersAdminNegativeTestJSON-229078574-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.974s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 575.920907] env[63088]: ERROR nova.compute.manager [None req-710306d6-422f-4cdd-91a1-33da212e5f03 tempest-ServersAdminNegativeTestJSON-229078574 tempest-ServersAdminNegativeTestJSON-229078574-project-member] [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f91b4261-8234-46b0-900f-271a33cad644, please check neutron logs for more information. [ 575.920907] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] Traceback (most recent call last): [ 575.920907] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 575.920907] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] self.driver.spawn(context, instance, image_meta, [ 575.920907] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 575.920907] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] self._vmops.spawn(context, instance, image_meta, injected_files, [ 575.920907] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 575.920907] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] vm_ref = self.build_virtual_machine(instance, [ 575.920907] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 575.920907] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] vif_infos = vmwarevif.get_vif_info(self._session, [ 575.920907] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 575.921228] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] for vif in network_info: [ 575.921228] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 575.921228] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] return self._sync_wrapper(fn, *args, **kwargs) [ 575.921228] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 575.921228] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] self.wait() [ 575.921228] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 575.921228] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] self[:] = self._gt.wait() [ 575.921228] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 575.921228] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] return self._exit_event.wait() [ 575.921228] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 575.921228] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] result = hub.switch() [ 575.921228] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 575.921228] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] return self.greenlet.switch() [ 575.921558] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 575.921558] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] result = function(*args, **kwargs) [ 575.921558] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 575.921558] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] return func(*args, **kwargs) [ 575.921558] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 575.921558] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] raise e [ 575.921558] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 575.921558] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] nwinfo = self.network_api.allocate_for_instance( [ 575.921558] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 575.921558] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] created_port_ids = self._update_ports_for_instance( [ 575.921558] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 575.921558] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] with excutils.save_and_reraise_exception(): [ 575.921558] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 575.921940] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] self.force_reraise() [ 575.921940] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 575.921940] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] raise self.value [ 575.921940] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 575.921940] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] updated_port = self._update_port( [ 575.921940] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 575.921940] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] _ensure_no_port_binding_failure(port) [ 575.921940] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 575.921940] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] raise exception.PortBindingFailed(port_id=port['id']) [ 575.921940] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] nova.exception.PortBindingFailed: Binding failed for port f91b4261-8234-46b0-900f-271a33cad644, please check neutron logs for more information. [ 575.921940] env[63088]: ERROR nova.compute.manager [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] [ 575.922250] env[63088]: DEBUG nova.compute.utils [None req-710306d6-422f-4cdd-91a1-33da212e5f03 tempest-ServersAdminNegativeTestJSON-229078574 tempest-ServersAdminNegativeTestJSON-229078574-project-member] [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] Binding failed for port f91b4261-8234-46b0-900f-271a33cad644, please check neutron logs for more information. {{(pid=63088) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 575.925157] env[63088]: DEBUG nova.compute.manager [None req-710306d6-422f-4cdd-91a1-33da212e5f03 tempest-ServersAdminNegativeTestJSON-229078574 tempest-ServersAdminNegativeTestJSON-229078574-project-member] [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] Build of instance 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52 was re-scheduled: Binding failed for port f91b4261-8234-46b0-900f-271a33cad644, please check neutron logs for more information. {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 575.925157] env[63088]: DEBUG nova.compute.manager [None req-710306d6-422f-4cdd-91a1-33da212e5f03 tempest-ServersAdminNegativeTestJSON-229078574 tempest-ServersAdminNegativeTestJSON-229078574-project-member] [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] Unplugging VIFs for instance {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 575.925157] env[63088]: DEBUG oslo_concurrency.lockutils [None req-710306d6-422f-4cdd-91a1-33da212e5f03 tempest-ServersAdminNegativeTestJSON-229078574 tempest-ServersAdminNegativeTestJSON-229078574-project-member] Acquiring lock "refresh_cache-4f487d47-daf7-4ef1-9e8b-8d8f870bfe52" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 575.925157] env[63088]: DEBUG oslo_concurrency.lockutils [None req-710306d6-422f-4cdd-91a1-33da212e5f03 tempest-ServersAdminNegativeTestJSON-229078574 tempest-ServersAdminNegativeTestJSON-229078574-project-member] Acquired lock "refresh_cache-4f487d47-daf7-4ef1-9e8b-8d8f870bfe52" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 575.925363] env[63088]: DEBUG nova.network.neutron [None req-710306d6-422f-4cdd-91a1-33da212e5f03 tempest-ServersAdminNegativeTestJSON-229078574 tempest-ServersAdminNegativeTestJSON-229078574-project-member] [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 575.925931] env[63088]: DEBUG oslo_concurrency.lockutils [None req-8a34ea86-0280-447d-8082-40a2090a8af5 tempest-ServerDiagnosticsTest-1360445823 tempest-ServerDiagnosticsTest-1360445823-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.620s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 576.001287] env[63088]: DEBUG oslo_vmware.api [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Task: {'id': task-1284783, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.049538} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 576.002374] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 576.003053] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-454ac76c-827d-4f0b-b621-253ebdc4822e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.009290] env[63088]: DEBUG oslo_vmware.api [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Waiting for the task: (returnval){ [ 576.009290] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52a3dc32-a23c-7f2d-4709-5e25b152e6e5" [ 576.009290] env[63088]: _type = "Task" [ 576.009290] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 576.022848] env[63088]: DEBUG oslo_vmware.api [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52a3dc32-a23c-7f2d-4709-5e25b152e6e5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 576.086403] env[63088]: INFO nova.compute.manager [None req-fa54bfad-a3c9-4ca3-8f38-8bc8499d1488 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] [instance: 05d2ad93-949e-42e3-85e9-eed9baaebe8c] Took 1.03 seconds to deallocate network for instance. [ 576.199176] env[63088]: DEBUG nova.network.neutron [req-cf482d53-8aca-4cbb-88da-85a2ba03e975 req-1439ab49-7837-4c3f-a8a5-1f897672cb51 service nova] [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 576.362963] env[63088]: DEBUG nova.network.neutron [req-cf482d53-8aca-4cbb-88da-85a2ba03e975 req-1439ab49-7837-4c3f-a8a5-1f897672cb51 service nova] [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 576.369230] env[63088]: DEBUG nova.compute.manager [None req-79832ead-48a0-4aad-a47e-8e38c2a687da tempest-ServersTestJSON-106593364 tempest-ServersTestJSON-106593364-project-member] [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 576.460114] env[63088]: DEBUG nova.network.neutron [None req-710306d6-422f-4cdd-91a1-33da212e5f03 tempest-ServersAdminNegativeTestJSON-229078574 tempest-ServersAdminNegativeTestJSON-229078574-project-member] [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 576.522266] env[63088]: DEBUG oslo_vmware.api [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52a3dc32-a23c-7f2d-4709-5e25b152e6e5, 'name': SearchDatastore_Task, 'duration_secs': 0.009036} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 576.525945] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 576.527629] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore2] c4777721-3f65-455f-9973-c1ed0732de34/c4777721-3f65-455f-9973-c1ed0732de34.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 576.528274] env[63088]: DEBUG oslo_concurrency.lockutils [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 576.528977] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 576.528977] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-be577f0c-1346-49c2-b656-0f906f5c1fda {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.530831] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1aa13a5b-6f0f-4908-bf95-9eb35419eae6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.538604] env[63088]: DEBUG oslo_vmware.api [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Waiting for the task: (returnval){ [ 576.538604] env[63088]: value = "task-1284784" [ 576.538604] env[63088]: _type = "Task" [ 576.538604] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 576.542354] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 576.542601] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 576.546978] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f7f1ef37-f014-4d1a-8075-90ca5f26b1a4 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.558892] env[63088]: DEBUG oslo_vmware.api [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Task: {'id': task-1284784, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 576.562992] env[63088]: DEBUG oslo_vmware.api [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Waiting for the task: (returnval){ [ 576.562992] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]525da99a-4da7-b1d8-d512-17c93f483a88" [ 576.562992] env[63088]: _type = "Task" [ 576.562992] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 576.564859] env[63088]: DEBUG nova.network.neutron [None req-710306d6-422f-4cdd-91a1-33da212e5f03 tempest-ServersAdminNegativeTestJSON-229078574 tempest-ServersAdminNegativeTestJSON-229078574-project-member] [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 576.577175] env[63088]: DEBUG oslo_vmware.api [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]525da99a-4da7-b1d8-d512-17c93f483a88, 'name': SearchDatastore_Task, 'duration_secs': 0.011465} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 576.578037] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6e4f54bd-a640-468d-a0da-9c461be0a1e4 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.587351] env[63088]: DEBUG oslo_vmware.api [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Waiting for the task: (returnval){ [ 576.587351] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52420081-2fef-bc6c-ee3e-1570eb525a28" [ 576.587351] env[63088]: _type = "Task" [ 576.587351] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 576.602026] env[63088]: DEBUG oslo_vmware.api [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52420081-2fef-bc6c-ee3e-1570eb525a28, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 576.795316] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-065281cb-cc46-4de7-b17e-2349e8fed309 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.808064] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dea72894-1d54-43ee-8ad9-728fd8f7e6c4 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.846356] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b1ed133-f7dd-442c-b305-138e1ea86a3b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.855797] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e934423-4a7f-4679-b78e-1fea58bb6be5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.870628] env[63088]: DEBUG oslo_concurrency.lockutils [req-cf482d53-8aca-4cbb-88da-85a2ba03e975 req-1439ab49-7837-4c3f-a8a5-1f897672cb51 service nova] Releasing lock "refresh_cache-44e39b2b-2bac-4a6e-9534-d2774d094931" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 576.871498] env[63088]: DEBUG nova.compute.manager [req-cf482d53-8aca-4cbb-88da-85a2ba03e975 req-1439ab49-7837-4c3f-a8a5-1f897672cb51 service nova] [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] Received event network-vif-deleted-27f3a8b8-19ea-48a7-8323-608f6af9414c {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 576.871498] env[63088]: DEBUG nova.compute.provider_tree [None req-8a34ea86-0280-447d-8082-40a2090a8af5 tempest-ServerDiagnosticsTest-1360445823 tempest-ServerDiagnosticsTest-1360445823-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 576.894229] env[63088]: DEBUG oslo_concurrency.lockutils [None req-79832ead-48a0-4aad-a47e-8e38c2a687da tempest-ServersTestJSON-106593364 tempest-ServersTestJSON-106593364-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 577.060439] env[63088]: DEBUG oslo_vmware.api [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Task: {'id': task-1284784, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 577.072653] env[63088]: DEBUG oslo_concurrency.lockutils [None req-710306d6-422f-4cdd-91a1-33da212e5f03 tempest-ServersAdminNegativeTestJSON-229078574 tempest-ServersAdminNegativeTestJSON-229078574-project-member] Releasing lock "refresh_cache-4f487d47-daf7-4ef1-9e8b-8d8f870bfe52" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 577.072947] env[63088]: DEBUG nova.compute.manager [None req-710306d6-422f-4cdd-91a1-33da212e5f03 tempest-ServersAdminNegativeTestJSON-229078574 tempest-ServersAdminNegativeTestJSON-229078574-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 577.073807] env[63088]: DEBUG nova.compute.manager [None req-710306d6-422f-4cdd-91a1-33da212e5f03 tempest-ServersAdminNegativeTestJSON-229078574 tempest-ServersAdminNegativeTestJSON-229078574-project-member] [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 577.073807] env[63088]: DEBUG nova.network.neutron [None req-710306d6-422f-4cdd-91a1-33da212e5f03 tempest-ServersAdminNegativeTestJSON-229078574 tempest-ServersAdminNegativeTestJSON-229078574-project-member] [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 577.097836] env[63088]: DEBUG oslo_vmware.api [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52420081-2fef-bc6c-ee3e-1570eb525a28, 'name': SearchDatastore_Task, 'duration_secs': 0.009775} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 577.098309] env[63088]: DEBUG oslo_concurrency.lockutils [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 577.098641] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore2] e8cfbce4-cfb2-463e-9f38-78a058dab6e1/e8cfbce4-cfb2-463e-9f38-78a058dab6e1.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 577.099406] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-15f7072a-8655-4353-b8f6-27f9a6049c5b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.109994] env[63088]: DEBUG oslo_vmware.api [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Waiting for the task: (returnval){ [ 577.109994] env[63088]: value = "task-1284785" [ 577.109994] env[63088]: _type = "Task" [ 577.109994] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 577.111080] env[63088]: DEBUG nova.network.neutron [None req-710306d6-422f-4cdd-91a1-33da212e5f03 tempest-ServersAdminNegativeTestJSON-229078574 tempest-ServersAdminNegativeTestJSON-229078574-project-member] [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 577.122039] env[63088]: DEBUG oslo_vmware.api [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Task: {'id': task-1284785, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 577.136343] env[63088]: INFO nova.scheduler.client.report [None req-fa54bfad-a3c9-4ca3-8f38-8bc8499d1488 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] Deleted allocations for instance 05d2ad93-949e-42e3-85e9-eed9baaebe8c [ 577.377248] env[63088]: DEBUG nova.scheduler.client.report [None req-8a34ea86-0280-447d-8082-40a2090a8af5 tempest-ServerDiagnosticsTest-1360445823 tempest-ServerDiagnosticsTest-1360445823-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 577.556146] env[63088]: DEBUG oslo_vmware.api [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Task: {'id': task-1284784, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.554757} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 577.556276] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore2] c4777721-3f65-455f-9973-c1ed0732de34/c4777721-3f65-455f-9973-c1ed0732de34.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 577.556504] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] [instance: c4777721-3f65-455f-9973-c1ed0732de34] Extending root virtual disk to 1048576 {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 577.557150] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-322a1d37-b17f-4c28-8019-450a43481f6f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.564045] env[63088]: DEBUG oslo_vmware.api [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Waiting for the task: (returnval){ [ 577.564045] env[63088]: value = "task-1284786" [ 577.564045] env[63088]: _type = "Task" [ 577.564045] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 577.574129] env[63088]: DEBUG oslo_vmware.api [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Task: {'id': task-1284786, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 577.621851] env[63088]: DEBUG nova.network.neutron [None req-710306d6-422f-4cdd-91a1-33da212e5f03 tempest-ServersAdminNegativeTestJSON-229078574 tempest-ServersAdminNegativeTestJSON-229078574-project-member] [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 577.630266] env[63088]: DEBUG oslo_vmware.api [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Task: {'id': task-1284785, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.453141} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 577.630266] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore2] e8cfbce4-cfb2-463e-9f38-78a058dab6e1/e8cfbce4-cfb2-463e-9f38-78a058dab6e1.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 577.630464] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Extending root virtual disk to 1048576 {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 577.630751] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c7f14659-b347-4eaa-a28a-a8e86802fbd8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.640708] env[63088]: DEBUG oslo_vmware.api [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Waiting for the task: (returnval){ [ 577.640708] env[63088]: value = "task-1284787" [ 577.640708] env[63088]: _type = "Task" [ 577.640708] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 577.648185] env[63088]: DEBUG oslo_vmware.api [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Task: {'id': task-1284787, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 577.649735] env[63088]: DEBUG oslo_concurrency.lockutils [None req-fa54bfad-a3c9-4ca3-8f38-8bc8499d1488 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] Lock "05d2ad93-949e-42e3-85e9-eed9baaebe8c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 33.434s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 577.885139] env[63088]: DEBUG oslo_concurrency.lockutils [None req-8a34ea86-0280-447d-8082-40a2090a8af5 tempest-ServerDiagnosticsTest-1360445823 tempest-ServerDiagnosticsTest-1360445823-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.959s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 577.886032] env[63088]: ERROR nova.compute.manager [None req-8a34ea86-0280-447d-8082-40a2090a8af5 tempest-ServerDiagnosticsTest-1360445823 tempest-ServerDiagnosticsTest-1360445823-project-member] [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 784e2da7-96ad-48a1-9570-f494c3f4b49a, please check neutron logs for more information. [ 577.886032] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] Traceback (most recent call last): [ 577.886032] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 577.886032] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] self.driver.spawn(context, instance, image_meta, [ 577.886032] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 577.886032] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] self._vmops.spawn(context, instance, image_meta, injected_files, [ 577.886032] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 577.886032] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] vm_ref = self.build_virtual_machine(instance, [ 577.886032] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 577.886032] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] vif_infos = vmwarevif.get_vif_info(self._session, [ 577.886032] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 577.886369] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] for vif in network_info: [ 577.886369] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 577.886369] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] return self._sync_wrapper(fn, *args, **kwargs) [ 577.886369] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 577.886369] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] self.wait() [ 577.886369] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 577.886369] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] self[:] = self._gt.wait() [ 577.886369] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 577.886369] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] return self._exit_event.wait() [ 577.886369] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 577.886369] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] result = hub.switch() [ 577.886369] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 577.886369] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] return self.greenlet.switch() [ 577.886755] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 577.886755] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] result = function(*args, **kwargs) [ 577.886755] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 577.886755] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] return func(*args, **kwargs) [ 577.886755] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 577.886755] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] raise e [ 577.886755] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 577.886755] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] nwinfo = self.network_api.allocate_for_instance( [ 577.886755] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 577.886755] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] created_port_ids = self._update_ports_for_instance( [ 577.886755] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 577.886755] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] with excutils.save_and_reraise_exception(): [ 577.886755] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 577.887195] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] self.force_reraise() [ 577.887195] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 577.887195] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] raise self.value [ 577.887195] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 577.887195] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] updated_port = self._update_port( [ 577.887195] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 577.887195] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] _ensure_no_port_binding_failure(port) [ 577.887195] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 577.887195] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] raise exception.PortBindingFailed(port_id=port['id']) [ 577.887195] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] nova.exception.PortBindingFailed: Binding failed for port 784e2da7-96ad-48a1-9570-f494c3f4b49a, please check neutron logs for more information. [ 577.887195] env[63088]: ERROR nova.compute.manager [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] [ 577.887524] env[63088]: DEBUG nova.compute.utils [None req-8a34ea86-0280-447d-8082-40a2090a8af5 tempest-ServerDiagnosticsTest-1360445823 tempest-ServerDiagnosticsTest-1360445823-project-member] [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] Binding failed for port 784e2da7-96ad-48a1-9570-f494c3f4b49a, please check neutron logs for more information. {{(pid=63088) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 577.889618] env[63088]: DEBUG oslo_concurrency.lockutils [None req-61dd383c-8cef-4a29-a464-69dfd03cb5ce tempest-ServerExternalEventsTest-515168927 tempest-ServerExternalEventsTest-515168927-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 10.017s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 577.893404] env[63088]: DEBUG nova.compute.manager [None req-8a34ea86-0280-447d-8082-40a2090a8af5 tempest-ServerDiagnosticsTest-1360445823 tempest-ServerDiagnosticsTest-1360445823-project-member] [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] Build of instance 2d7bf024-afb5-4d33-9c5e-06633167ac77 was re-scheduled: Binding failed for port 784e2da7-96ad-48a1-9570-f494c3f4b49a, please check neutron logs for more information. {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 577.894095] env[63088]: DEBUG nova.compute.manager [None req-8a34ea86-0280-447d-8082-40a2090a8af5 tempest-ServerDiagnosticsTest-1360445823 tempest-ServerDiagnosticsTest-1360445823-project-member] [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] Unplugging VIFs for instance {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 577.899161] env[63088]: DEBUG oslo_concurrency.lockutils [None req-8a34ea86-0280-447d-8082-40a2090a8af5 tempest-ServerDiagnosticsTest-1360445823 tempest-ServerDiagnosticsTest-1360445823-project-member] Acquiring lock "refresh_cache-2d7bf024-afb5-4d33-9c5e-06633167ac77" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 577.899161] env[63088]: DEBUG oslo_concurrency.lockutils [None req-8a34ea86-0280-447d-8082-40a2090a8af5 tempest-ServerDiagnosticsTest-1360445823 tempest-ServerDiagnosticsTest-1360445823-project-member] Acquired lock "refresh_cache-2d7bf024-afb5-4d33-9c5e-06633167ac77" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 577.899161] env[63088]: DEBUG nova.network.neutron [None req-8a34ea86-0280-447d-8082-40a2090a8af5 tempest-ServerDiagnosticsTest-1360445823 tempest-ServerDiagnosticsTest-1360445823-project-member] [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 578.077410] env[63088]: DEBUG oslo_vmware.api [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Task: {'id': task-1284786, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068335} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 578.077689] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] [instance: c4777721-3f65-455f-9973-c1ed0732de34] Extended root virtual disk {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 578.078526] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e30fc301-a56d-4c98-898b-46cc51c1c319 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.105483] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] [instance: c4777721-3f65-455f-9973-c1ed0732de34] Reconfiguring VM instance instance-00000009 to attach disk [datastore2] c4777721-3f65-455f-9973-c1ed0732de34/c4777721-3f65-455f-9973-c1ed0732de34.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 578.106481] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-43851ab4-d509-4860-8266-1bcc1b2ee685 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.127287] env[63088]: INFO nova.compute.manager [None req-710306d6-422f-4cdd-91a1-33da212e5f03 tempest-ServersAdminNegativeTestJSON-229078574 tempest-ServersAdminNegativeTestJSON-229078574-project-member] [instance: 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52] Took 1.05 seconds to deallocate network for instance. [ 578.141147] env[63088]: DEBUG oslo_vmware.api [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Waiting for the task: (returnval){ [ 578.141147] env[63088]: value = "task-1284788" [ 578.141147] env[63088]: _type = "Task" [ 578.141147] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 578.153097] env[63088]: DEBUG nova.compute.manager [None req-a9773ab2-f8b9-4102-8e41-5ee881a8637e tempest-FloatingIPsAssociationNegativeTestJSON-189613961 tempest-FloatingIPsAssociationNegativeTestJSON-189613961-project-member] [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 578.159764] env[63088]: DEBUG oslo_vmware.api [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Task: {'id': task-1284787, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062974} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 578.160335] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Extended root virtual disk {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 578.161215] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65c8acf6-8ceb-4852-8d39-e7b6e4142698 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.169765] env[63088]: DEBUG oslo_vmware.api [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Task: {'id': task-1284788, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 578.188790] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Reconfiguring VM instance instance-00000005 to attach disk [datastore2] e8cfbce4-cfb2-463e-9f38-78a058dab6e1/e8cfbce4-cfb2-463e-9f38-78a058dab6e1.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 578.189264] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1f0ab7a1-e966-4cf2-a485-e1817cdf827a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.210152] env[63088]: DEBUG oslo_vmware.api [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Waiting for the task: (returnval){ [ 578.210152] env[63088]: value = "task-1284789" [ 578.210152] env[63088]: _type = "Task" [ 578.210152] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 578.224256] env[63088]: DEBUG oslo_vmware.api [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Task: {'id': task-1284789, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 578.465048] env[63088]: DEBUG nova.network.neutron [None req-8a34ea86-0280-447d-8082-40a2090a8af5 tempest-ServerDiagnosticsTest-1360445823 tempest-ServerDiagnosticsTest-1360445823-project-member] [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 578.664217] env[63088]: DEBUG oslo_vmware.api [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Task: {'id': task-1284788, 'name': ReconfigVM_Task, 'duration_secs': 0.297971} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 578.668380] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] [instance: c4777721-3f65-455f-9973-c1ed0732de34] Reconfigured VM instance instance-00000009 to attach disk [datastore2] c4777721-3f65-455f-9973-c1ed0732de34/c4777721-3f65-455f-9973-c1ed0732de34.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 578.668380] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-feab48b9-e8cd-42d4-ba2b-2a46062dc93f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.678648] env[63088]: DEBUG oslo_vmware.api [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Waiting for the task: (returnval){ [ 578.678648] env[63088]: value = "task-1284790" [ 578.678648] env[63088]: _type = "Task" [ 578.678648] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 578.691998] env[63088]: DEBUG nova.network.neutron [None req-8a34ea86-0280-447d-8082-40a2090a8af5 tempest-ServerDiagnosticsTest-1360445823 tempest-ServerDiagnosticsTest-1360445823-project-member] [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 578.694078] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a9773ab2-f8b9-4102-8e41-5ee881a8637e tempest-FloatingIPsAssociationNegativeTestJSON-189613961 tempest-FloatingIPsAssociationNegativeTestJSON-189613961-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 578.701130] env[63088]: DEBUG oslo_vmware.api [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Task: {'id': task-1284790, 'name': Rename_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 578.733464] env[63088]: DEBUG oslo_vmware.api [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Task: {'id': task-1284789, 'name': ReconfigVM_Task, 'duration_secs': 0.289916} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 578.736655] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Reconfigured VM instance instance-00000005 to attach disk [datastore2] e8cfbce4-cfb2-463e-9f38-78a058dab6e1/e8cfbce4-cfb2-463e-9f38-78a058dab6e1.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 578.737620] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3192402e-bc81-4d67-9829-feb6a071c2e0 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.745906] env[63088]: DEBUG oslo_vmware.api [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Waiting for the task: (returnval){ [ 578.745906] env[63088]: value = "task-1284791" [ 578.745906] env[63088]: _type = "Task" [ 578.745906] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 578.764345] env[63088]: DEBUG oslo_vmware.api [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Task: {'id': task-1284791, 'name': Rename_Task} progress is 10%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 578.800711] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e2e2fa6-5e58-43a1-82fc-a066a263ac63 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.808735] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d1cb247-2e4e-40af-b6b0-5f34c955232f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.854489] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efcf84b0-76da-4451-8e7b-a6d4f64c3591 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.868379] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1199008-09fb-441d-875b-a963db3d4300 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.884100] env[63088]: DEBUG nova.compute.provider_tree [None req-61dd383c-8cef-4a29-a464-69dfd03cb5ce tempest-ServerExternalEventsTest-515168927 tempest-ServerExternalEventsTest-515168927-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 579.185634] env[63088]: INFO nova.scheduler.client.report [None req-710306d6-422f-4cdd-91a1-33da212e5f03 tempest-ServersAdminNegativeTestJSON-229078574 tempest-ServersAdminNegativeTestJSON-229078574-project-member] Deleted allocations for instance 4f487d47-daf7-4ef1-9e8b-8d8f870bfe52 [ 579.203182] env[63088]: DEBUG oslo_concurrency.lockutils [None req-8a34ea86-0280-447d-8082-40a2090a8af5 tempest-ServerDiagnosticsTest-1360445823 tempest-ServerDiagnosticsTest-1360445823-project-member] Releasing lock "refresh_cache-2d7bf024-afb5-4d33-9c5e-06633167ac77" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 579.203376] env[63088]: DEBUG nova.compute.manager [None req-8a34ea86-0280-447d-8082-40a2090a8af5 tempest-ServerDiagnosticsTest-1360445823 tempest-ServerDiagnosticsTest-1360445823-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 579.203543] env[63088]: DEBUG nova.compute.manager [None req-8a34ea86-0280-447d-8082-40a2090a8af5 tempest-ServerDiagnosticsTest-1360445823 tempest-ServerDiagnosticsTest-1360445823-project-member] [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 579.203687] env[63088]: DEBUG nova.network.neutron [None req-8a34ea86-0280-447d-8082-40a2090a8af5 tempest-ServerDiagnosticsTest-1360445823 tempest-ServerDiagnosticsTest-1360445823-project-member] [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 579.205421] env[63088]: DEBUG oslo_vmware.api [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Task: {'id': task-1284790, 'name': Rename_Task, 'duration_secs': 0.141071} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 579.206720] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] [instance: c4777721-3f65-455f-9973-c1ed0732de34] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 579.206720] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-32a8f93b-2656-4080-ab3c-b185c6e2caf5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.217369] env[63088]: DEBUG oslo_vmware.api [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Waiting for the task: (returnval){ [ 579.217369] env[63088]: value = "task-1284792" [ 579.217369] env[63088]: _type = "Task" [ 579.217369] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 579.223799] env[63088]: DEBUG oslo_vmware.api [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Task: {'id': task-1284792, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 579.230364] env[63088]: DEBUG nova.network.neutron [None req-8a34ea86-0280-447d-8082-40a2090a8af5 tempest-ServerDiagnosticsTest-1360445823 tempest-ServerDiagnosticsTest-1360445823-project-member] [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 579.257733] env[63088]: DEBUG oslo_vmware.api [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Task: {'id': task-1284791, 'name': Rename_Task, 'duration_secs': 0.172546} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 579.258096] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 579.258303] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cf0d5dfd-4c7a-4132-a659-68fb775ebf6b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.266441] env[63088]: DEBUG oslo_vmware.api [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Waiting for the task: (returnval){ [ 579.266441] env[63088]: value = "task-1284793" [ 579.266441] env[63088]: _type = "Task" [ 579.266441] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 579.279412] env[63088]: DEBUG oslo_vmware.api [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Task: {'id': task-1284793, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 579.388696] env[63088]: DEBUG nova.scheduler.client.report [None req-61dd383c-8cef-4a29-a464-69dfd03cb5ce tempest-ServerExternalEventsTest-515168927 tempest-ServerExternalEventsTest-515168927-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 579.392060] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f81b332d-72eb-4149-a6a2-3c8a9c9bec53 tempest-ServersNegativeTestJSON-1008333314 tempest-ServersNegativeTestJSON-1008333314-project-member] Acquiring lock "5b778247-1434-447e-8ff1-7359bb9e8ea2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 579.392293] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f81b332d-72eb-4149-a6a2-3c8a9c9bec53 tempest-ServersNegativeTestJSON-1008333314 tempest-ServersNegativeTestJSON-1008333314-project-member] Lock "5b778247-1434-447e-8ff1-7359bb9e8ea2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 579.709253] env[63088]: DEBUG oslo_concurrency.lockutils [None req-710306d6-422f-4cdd-91a1-33da212e5f03 tempest-ServersAdminNegativeTestJSON-229078574 tempest-ServersAdminNegativeTestJSON-229078574-project-member] Lock "4f487d47-daf7-4ef1-9e8b-8d8f870bfe52" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.197s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 579.726256] env[63088]: DEBUG oslo_vmware.api [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Task: {'id': task-1284792, 'name': PowerOnVM_Task, 'duration_secs': 0.446021} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 579.726397] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] [instance: c4777721-3f65-455f-9973-c1ed0732de34] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 579.726560] env[63088]: INFO nova.compute.manager [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] [instance: c4777721-3f65-455f-9973-c1ed0732de34] Took 10.14 seconds to spawn the instance on the hypervisor. [ 579.726738] env[63088]: DEBUG nova.compute.manager [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] [instance: c4777721-3f65-455f-9973-c1ed0732de34] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 579.727543] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0634bce8-1cb5-4653-987d-1686149b8c33 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.734847] env[63088]: DEBUG nova.network.neutron [None req-8a34ea86-0280-447d-8082-40a2090a8af5 tempest-ServerDiagnosticsTest-1360445823 tempest-ServerDiagnosticsTest-1360445823-project-member] [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 579.778558] env[63088]: DEBUG oslo_vmware.api [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Task: {'id': task-1284793, 'name': PowerOnVM_Task, 'duration_secs': 0.506457} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 579.778933] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 579.779268] env[63088]: DEBUG nova.compute.manager [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 579.779928] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f67f54b6-aed0-4cba-b20f-800237c5d5c1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.893941] env[63088]: DEBUG oslo_concurrency.lockutils [None req-61dd383c-8cef-4a29-a464-69dfd03cb5ce tempest-ServerExternalEventsTest-515168927 tempest-ServerExternalEventsTest-515168927-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.004s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 579.894616] env[63088]: ERROR nova.compute.manager [None req-61dd383c-8cef-4a29-a464-69dfd03cb5ce tempest-ServerExternalEventsTest-515168927 tempest-ServerExternalEventsTest-515168927-project-member] [instance: 173fccf1-f258-4f0c-a968-466a27406c55] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 06dbc948-3bcb-4bb1-9083-7a5a312da444, please check neutron logs for more information. [ 579.894616] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] Traceback (most recent call last): [ 579.894616] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 579.894616] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] self.driver.spawn(context, instance, image_meta, [ 579.894616] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 579.894616] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] self._vmops.spawn(context, instance, image_meta, injected_files, [ 579.894616] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 579.894616] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] vm_ref = self.build_virtual_machine(instance, [ 579.894616] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 579.894616] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] vif_infos = vmwarevif.get_vif_info(self._session, [ 579.894616] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 579.895088] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] for vif in network_info: [ 579.895088] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 579.895088] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] return self._sync_wrapper(fn, *args, **kwargs) [ 579.895088] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 579.895088] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] self.wait() [ 579.895088] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 579.895088] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] self[:] = self._gt.wait() [ 579.895088] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 579.895088] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] return self._exit_event.wait() [ 579.895088] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 579.895088] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] result = hub.switch() [ 579.895088] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 579.895088] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] return self.greenlet.switch() [ 579.895488] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 579.895488] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] result = function(*args, **kwargs) [ 579.895488] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 579.895488] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] return func(*args, **kwargs) [ 579.895488] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 579.895488] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] raise e [ 579.895488] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 579.895488] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] nwinfo = self.network_api.allocate_for_instance( [ 579.895488] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 579.895488] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] created_port_ids = self._update_ports_for_instance( [ 579.895488] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 579.895488] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] with excutils.save_and_reraise_exception(): [ 579.895488] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 579.895883] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] self.force_reraise() [ 579.895883] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 579.895883] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] raise self.value [ 579.895883] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 579.895883] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] updated_port = self._update_port( [ 579.895883] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 579.895883] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] _ensure_no_port_binding_failure(port) [ 579.895883] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 579.895883] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] raise exception.PortBindingFailed(port_id=port['id']) [ 579.895883] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] nova.exception.PortBindingFailed: Binding failed for port 06dbc948-3bcb-4bb1-9083-7a5a312da444, please check neutron logs for more information. [ 579.895883] env[63088]: ERROR nova.compute.manager [instance: 173fccf1-f258-4f0c-a968-466a27406c55] [ 579.896229] env[63088]: DEBUG nova.compute.utils [None req-61dd383c-8cef-4a29-a464-69dfd03cb5ce tempest-ServerExternalEventsTest-515168927 tempest-ServerExternalEventsTest-515168927-project-member] [instance: 173fccf1-f258-4f0c-a968-466a27406c55] Binding failed for port 06dbc948-3bcb-4bb1-9083-7a5a312da444, please check neutron logs for more information. {{(pid=63088) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 579.896609] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d1272b04-ee5c-4e9e-b0bf-6ca9d0629aae tempest-VolumesAssistedSnapshotsTest-1004656420 tempest-VolumesAssistedSnapshotsTest-1004656420-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.392s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 579.898342] env[63088]: INFO nova.compute.claims [None req-d1272b04-ee5c-4e9e-b0bf-6ca9d0629aae tempest-VolumesAssistedSnapshotsTest-1004656420 tempest-VolumesAssistedSnapshotsTest-1004656420-project-member] [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 579.901626] env[63088]: DEBUG nova.compute.manager [None req-61dd383c-8cef-4a29-a464-69dfd03cb5ce tempest-ServerExternalEventsTest-515168927 tempest-ServerExternalEventsTest-515168927-project-member] [instance: 173fccf1-f258-4f0c-a968-466a27406c55] Build of instance 173fccf1-f258-4f0c-a968-466a27406c55 was re-scheduled: Binding failed for port 06dbc948-3bcb-4bb1-9083-7a5a312da444, please check neutron logs for more information. {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 579.902187] env[63088]: DEBUG nova.compute.manager [None req-61dd383c-8cef-4a29-a464-69dfd03cb5ce tempest-ServerExternalEventsTest-515168927 tempest-ServerExternalEventsTest-515168927-project-member] [instance: 173fccf1-f258-4f0c-a968-466a27406c55] Unplugging VIFs for instance {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 579.902329] env[63088]: DEBUG oslo_concurrency.lockutils [None req-61dd383c-8cef-4a29-a464-69dfd03cb5ce tempest-ServerExternalEventsTest-515168927 tempest-ServerExternalEventsTest-515168927-project-member] Acquiring lock "refresh_cache-173fccf1-f258-4f0c-a968-466a27406c55" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 579.902474] env[63088]: DEBUG oslo_concurrency.lockutils [None req-61dd383c-8cef-4a29-a464-69dfd03cb5ce tempest-ServerExternalEventsTest-515168927 tempest-ServerExternalEventsTest-515168927-project-member] Acquired lock "refresh_cache-173fccf1-f258-4f0c-a968-466a27406c55" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 579.902649] env[63088]: DEBUG nova.network.neutron [None req-61dd383c-8cef-4a29-a464-69dfd03cb5ce tempest-ServerExternalEventsTest-515168927 tempest-ServerExternalEventsTest-515168927-project-member] [instance: 173fccf1-f258-4f0c-a968-466a27406c55] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 580.211690] env[63088]: DEBUG nova.compute.manager [None req-bd71c985-753f-4eb9-b297-7d458ea844aa tempest-ImagesOneServerNegativeTestJSON-1975365234 tempest-ImagesOneServerNegativeTestJSON-1975365234-project-member] [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 580.245029] env[63088]: INFO nova.compute.manager [None req-8a34ea86-0280-447d-8082-40a2090a8af5 tempest-ServerDiagnosticsTest-1360445823 tempest-ServerDiagnosticsTest-1360445823-project-member] [instance: 2d7bf024-afb5-4d33-9c5e-06633167ac77] Took 1.04 seconds to deallocate network for instance. [ 580.266414] env[63088]: INFO nova.compute.manager [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] [instance: c4777721-3f65-455f-9973-c1ed0732de34] Took 28.17 seconds to build instance. [ 580.302360] env[63088]: DEBUG oslo_concurrency.lockutils [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 580.461922] env[63088]: DEBUG nova.network.neutron [None req-61dd383c-8cef-4a29-a464-69dfd03cb5ce tempest-ServerExternalEventsTest-515168927 tempest-ServerExternalEventsTest-515168927-project-member] [instance: 173fccf1-f258-4f0c-a968-466a27406c55] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 580.550246] env[63088]: DEBUG nova.network.neutron [None req-61dd383c-8cef-4a29-a464-69dfd03cb5ce tempest-ServerExternalEventsTest-515168927 tempest-ServerExternalEventsTest-515168927-project-member] [instance: 173fccf1-f258-4f0c-a968-466a27406c55] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 580.746335] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bd71c985-753f-4eb9-b297-7d458ea844aa tempest-ImagesOneServerNegativeTestJSON-1975365234 tempest-ImagesOneServerNegativeTestJSON-1975365234-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 580.770550] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a9c0006e-0b53-492f-acd9-711c424e01cd tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Lock "c4777721-3f65-455f-9973-c1ed0732de34" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.678s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 580.812506] env[63088]: DEBUG nova.compute.manager [None req-1a3929a0-3bc6-4052-9624-82762b9682ba tempest-ServerDiagnosticsV248Test-626672433 tempest-ServerDiagnosticsV248Test-626672433-project-admin] [instance: c4777721-3f65-455f-9973-c1ed0732de34] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 580.813516] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad6923d2-d26d-4d6d-8f31-55d57014ec59 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.822713] env[63088]: INFO nova.compute.manager [None req-1a3929a0-3bc6-4052-9624-82762b9682ba tempest-ServerDiagnosticsV248Test-626672433 tempest-ServerDiagnosticsV248Test-626672433-project-admin] [instance: c4777721-3f65-455f-9973-c1ed0732de34] Retrieving diagnostics [ 580.823629] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56286525-7382-4226-81d0-6f507bb3f7b0 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.054389] env[63088]: DEBUG oslo_concurrency.lockutils [None req-61dd383c-8cef-4a29-a464-69dfd03cb5ce tempest-ServerExternalEventsTest-515168927 tempest-ServerExternalEventsTest-515168927-project-member] Releasing lock "refresh_cache-173fccf1-f258-4f0c-a968-466a27406c55" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 581.054626] env[63088]: DEBUG nova.compute.manager [None req-61dd383c-8cef-4a29-a464-69dfd03cb5ce tempest-ServerExternalEventsTest-515168927 tempest-ServerExternalEventsTest-515168927-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 581.054792] env[63088]: DEBUG nova.compute.manager [None req-61dd383c-8cef-4a29-a464-69dfd03cb5ce tempest-ServerExternalEventsTest-515168927 tempest-ServerExternalEventsTest-515168927-project-member] [instance: 173fccf1-f258-4f0c-a968-466a27406c55] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 581.054968] env[63088]: DEBUG nova.network.neutron [None req-61dd383c-8cef-4a29-a464-69dfd03cb5ce tempest-ServerExternalEventsTest-515168927 tempest-ServerExternalEventsTest-515168927-project-member] [instance: 173fccf1-f258-4f0c-a968-466a27406c55] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 581.092020] env[63088]: DEBUG nova.network.neutron [None req-61dd383c-8cef-4a29-a464-69dfd03cb5ce tempest-ServerExternalEventsTest-515168927 tempest-ServerExternalEventsTest-515168927-project-member] [instance: 173fccf1-f258-4f0c-a968-466a27406c55] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 581.219445] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc143b8f-c88e-4af9-b027-2e522179c096 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.228321] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86bd93b3-df81-4f31-a27c-dedc73f7869e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.281497] env[63088]: DEBUG nova.compute.manager [None req-83cac04a-fc75-40e5-8cdf-832028033c8d tempest-ServersWithSpecificFlavorTestJSON-501634894 tempest-ServersWithSpecificFlavorTestJSON-501634894-project-member] [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 581.284812] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26dbbbc4-bb02-4798-b214-b1f207f292e4 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.293317] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73d60f3a-8757-44e2-8bd3-944f4462555a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.310750] env[63088]: DEBUG nova.compute.provider_tree [None req-d1272b04-ee5c-4e9e-b0bf-6ca9d0629aae tempest-VolumesAssistedSnapshotsTest-1004656420 tempest-VolumesAssistedSnapshotsTest-1004656420-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 581.312653] env[63088]: INFO nova.scheduler.client.report [None req-8a34ea86-0280-447d-8082-40a2090a8af5 tempest-ServerDiagnosticsTest-1360445823 tempest-ServerDiagnosticsTest-1360445823-project-member] Deleted allocations for instance 2d7bf024-afb5-4d33-9c5e-06633167ac77 [ 581.595152] env[63088]: DEBUG nova.network.neutron [None req-61dd383c-8cef-4a29-a464-69dfd03cb5ce tempest-ServerExternalEventsTest-515168927 tempest-ServerExternalEventsTest-515168927-project-member] [instance: 173fccf1-f258-4f0c-a968-466a27406c55] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 581.806626] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a5e283fa-bcc1-4186-a4e3-1ae9f8d221af tempest-ServerMetadataTestJSON-859565163 tempest-ServerMetadataTestJSON-859565163-project-member] Acquiring lock "966fe303-09b9-449f-8bda-4f6cda135f7a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 581.806626] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a5e283fa-bcc1-4186-a4e3-1ae9f8d221af tempest-ServerMetadataTestJSON-859565163 tempest-ServerMetadataTestJSON-859565163-project-member] Lock "966fe303-09b9-449f-8bda-4f6cda135f7a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 581.812226] env[63088]: DEBUG oslo_concurrency.lockutils [None req-83cac04a-fc75-40e5-8cdf-832028033c8d tempest-ServersWithSpecificFlavorTestJSON-501634894 tempest-ServersWithSpecificFlavorTestJSON-501634894-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 581.819356] env[63088]: DEBUG nova.scheduler.client.report [None req-d1272b04-ee5c-4e9e-b0bf-6ca9d0629aae tempest-VolumesAssistedSnapshotsTest-1004656420 tempest-VolumesAssistedSnapshotsTest-1004656420-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 581.823698] env[63088]: DEBUG oslo_concurrency.lockutils [None req-8a34ea86-0280-447d-8082-40a2090a8af5 tempest-ServerDiagnosticsTest-1360445823 tempest-ServerDiagnosticsTest-1360445823-project-member] Lock "2d7bf024-afb5-4d33-9c5e-06633167ac77" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 37.065s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 582.101248] env[63088]: INFO nova.compute.manager [None req-61dd383c-8cef-4a29-a464-69dfd03cb5ce tempest-ServerExternalEventsTest-515168927 tempest-ServerExternalEventsTest-515168927-project-member] [instance: 173fccf1-f258-4f0c-a968-466a27406c55] Took 1.05 seconds to deallocate network for instance. [ 582.326254] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d1272b04-ee5c-4e9e-b0bf-6ca9d0629aae tempest-VolumesAssistedSnapshotsTest-1004656420 tempest-VolumesAssistedSnapshotsTest-1004656420-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.429s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 582.327080] env[63088]: DEBUG nova.compute.manager [None req-d1272b04-ee5c-4e9e-b0bf-6ca9d0629aae tempest-VolumesAssistedSnapshotsTest-1004656420 tempest-VolumesAssistedSnapshotsTest-1004656420-project-member] [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 582.332848] env[63088]: DEBUG nova.compute.manager [None req-acfde423-bb82-457b-a17c-3c463abca710 tempest-ServersTestFqdnHostnames-1568147653 tempest-ServersTestFqdnHostnames-1568147653-project-member] [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 582.342565] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 12.276s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 582.684658] env[63088]: DEBUG oslo_concurrency.lockutils [None req-224c474a-3c21-4a3f-9526-5bb4e3ba7d78 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] Acquiring lock "62bd4013-f007-4f69-b037-14bedf1e63cf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 582.685376] env[63088]: DEBUG oslo_concurrency.lockutils [None req-224c474a-3c21-4a3f-9526-5bb4e3ba7d78 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] Lock "62bd4013-f007-4f69-b037-14bedf1e63cf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 582.846247] env[63088]: DEBUG nova.compute.utils [None req-d1272b04-ee5c-4e9e-b0bf-6ca9d0629aae tempest-VolumesAssistedSnapshotsTest-1004656420 tempest-VolumesAssistedSnapshotsTest-1004656420-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 582.853632] env[63088]: DEBUG nova.compute.manager [None req-d1272b04-ee5c-4e9e-b0bf-6ca9d0629aae tempest-VolumesAssistedSnapshotsTest-1004656420 tempest-VolumesAssistedSnapshotsTest-1004656420-project-member] [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 582.854262] env[63088]: DEBUG nova.network.neutron [None req-d1272b04-ee5c-4e9e-b0bf-6ca9d0629aae tempest-VolumesAssistedSnapshotsTest-1004656420 tempest-VolumesAssistedSnapshotsTest-1004656420-project-member] [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 582.879871] env[63088]: DEBUG oslo_concurrency.lockutils [None req-acfde423-bb82-457b-a17c-3c463abca710 tempest-ServersTestFqdnHostnames-1568147653 tempest-ServersTestFqdnHostnames-1568147653-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 582.954367] env[63088]: DEBUG oslo_concurrency.lockutils [None req-cab9e1ae-5fd3-4816-8684-8d11c94f36d2 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] Acquiring lock "a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 582.954647] env[63088]: DEBUG oslo_concurrency.lockutils [None req-cab9e1ae-5fd3-4816-8684-8d11c94f36d2 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] Lock "a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 582.982429] env[63088]: DEBUG nova.policy [None req-d1272b04-ee5c-4e9e-b0bf-6ca9d0629aae tempest-VolumesAssistedSnapshotsTest-1004656420 tempest-VolumesAssistedSnapshotsTest-1004656420-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '91e80fe34258409784d453c9fb53c3bc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bc1d0f5436f14646808db8e6255c13df', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 583.153259] env[63088]: INFO nova.scheduler.client.report [None req-61dd383c-8cef-4a29-a464-69dfd03cb5ce tempest-ServerExternalEventsTest-515168927 tempest-ServerExternalEventsTest-515168927-project-member] Deleted allocations for instance 173fccf1-f258-4f0c-a968-466a27406c55 [ 583.355714] env[63088]: DEBUG nova.compute.manager [None req-d1272b04-ee5c-4e9e-b0bf-6ca9d0629aae tempest-VolumesAssistedSnapshotsTest-1004656420 tempest-VolumesAssistedSnapshotsTest-1004656420-project-member] [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 583.386101] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance e8cfbce4-cfb2-463e-9f38-78a058dab6e1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 583.386263] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance 5c3f72e3-d4f4-4c34-92a3-075536b85c99 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 583.386385] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance 4d5293a0-6945-4731-afb6-996a142c8447 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 583.386519] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance c4777721-3f65-455f-9973-c1ed0732de34 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 583.386619] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance 44e39b2b-2bac-4a6e-9534-d2774d094931 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 583.386727] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance d2c07175-b635-4900-9dab-9767e5f6bce4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 583.667663] env[63088]: DEBUG oslo_concurrency.lockutils [None req-61dd383c-8cef-4a29-a464-69dfd03cb5ce tempest-ServerExternalEventsTest-515168927 tempest-ServerExternalEventsTest-515168927-project-member] Lock "173fccf1-f258-4f0c-a968-466a27406c55" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 37.108s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 583.681025] env[63088]: DEBUG nova.network.neutron [None req-d1272b04-ee5c-4e9e-b0bf-6ca9d0629aae tempest-VolumesAssistedSnapshotsTest-1004656420 tempest-VolumesAssistedSnapshotsTest-1004656420-project-member] [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] Successfully created port: e0dd550d-f2d8-47df-9706-a8988ec41e69 {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 583.818961] env[63088]: INFO nova.compute.manager [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Rebuilding instance [ 583.872991] env[63088]: DEBUG nova.compute.manager [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 583.876022] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5a6b013-a2a4-4c23-94f8-dd7a604baba1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.893665] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance 34c01c2c-519b-44a1-aba2-c26e2c0b4737 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 584.173946] env[63088]: DEBUG nova.compute.manager [None req-f5f50761-0430-407c-84cc-7b2c9cf8683b tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 584.334350] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b7ea81d3-fe92-4160-8c6b-de9afb157d35 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] Acquiring lock "945b5a77-96b9-4565-94a8-6f1a89dc3bbf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 584.334602] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b7ea81d3-fe92-4160-8c6b-de9afb157d35 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] Lock "945b5a77-96b9-4565-94a8-6f1a89dc3bbf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.004s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 584.364966] env[63088]: DEBUG nova.compute.manager [None req-d1272b04-ee5c-4e9e-b0bf-6ca9d0629aae tempest-VolumesAssistedSnapshotsTest-1004656420 tempest-VolumesAssistedSnapshotsTest-1004656420-project-member] [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 584.387606] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 584.388022] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-216ed5f7-ad59-424f-9dd1-af64e6ffde21 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.401776] env[63088]: DEBUG nova.virt.hardware [None req-d1272b04-ee5c-4e9e-b0bf-6ca9d0629aae tempest-VolumesAssistedSnapshotsTest-1004656420 tempest-VolumesAssistedSnapshotsTest-1004656420-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 584.402343] env[63088]: DEBUG nova.virt.hardware [None req-d1272b04-ee5c-4e9e-b0bf-6ca9d0629aae tempest-VolumesAssistedSnapshotsTest-1004656420 tempest-VolumesAssistedSnapshotsTest-1004656420-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 584.402977] env[63088]: DEBUG nova.virt.hardware [None req-d1272b04-ee5c-4e9e-b0bf-6ca9d0629aae tempest-VolumesAssistedSnapshotsTest-1004656420 tempest-VolumesAssistedSnapshotsTest-1004656420-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 584.403275] env[63088]: DEBUG nova.virt.hardware [None req-d1272b04-ee5c-4e9e-b0bf-6ca9d0629aae tempest-VolumesAssistedSnapshotsTest-1004656420 tempest-VolumesAssistedSnapshotsTest-1004656420-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 584.403572] env[63088]: DEBUG nova.virt.hardware [None req-d1272b04-ee5c-4e9e-b0bf-6ca9d0629aae tempest-VolumesAssistedSnapshotsTest-1004656420 tempest-VolumesAssistedSnapshotsTest-1004656420-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 584.403804] env[63088]: DEBUG nova.virt.hardware [None req-d1272b04-ee5c-4e9e-b0bf-6ca9d0629aae tempest-VolumesAssistedSnapshotsTest-1004656420 tempest-VolumesAssistedSnapshotsTest-1004656420-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 584.404158] env[63088]: DEBUG nova.virt.hardware [None req-d1272b04-ee5c-4e9e-b0bf-6ca9d0629aae tempest-VolumesAssistedSnapshotsTest-1004656420 tempest-VolumesAssistedSnapshotsTest-1004656420-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 584.405020] env[63088]: DEBUG nova.virt.hardware [None req-d1272b04-ee5c-4e9e-b0bf-6ca9d0629aae tempest-VolumesAssistedSnapshotsTest-1004656420 tempest-VolumesAssistedSnapshotsTest-1004656420-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 584.405020] env[63088]: DEBUG nova.virt.hardware [None req-d1272b04-ee5c-4e9e-b0bf-6ca9d0629aae tempest-VolumesAssistedSnapshotsTest-1004656420 tempest-VolumesAssistedSnapshotsTest-1004656420-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 584.405020] env[63088]: DEBUG nova.virt.hardware [None req-d1272b04-ee5c-4e9e-b0bf-6ca9d0629aae tempest-VolumesAssistedSnapshotsTest-1004656420 tempest-VolumesAssistedSnapshotsTest-1004656420-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 584.405020] env[63088]: DEBUG nova.virt.hardware [None req-d1272b04-ee5c-4e9e-b0bf-6ca9d0629aae tempest-VolumesAssistedSnapshotsTest-1004656420 tempest-VolumesAssistedSnapshotsTest-1004656420-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 584.405715] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 584.408881] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98f2bc7b-879d-4757-ab30-84ca25a2b81c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.418678] env[63088]: DEBUG oslo_vmware.api [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] Waiting for the task: (returnval){ [ 584.418678] env[63088]: value = "task-1284794" [ 584.418678] env[63088]: _type = "Task" [ 584.418678] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 584.433938] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-208b2b27-6d1a-45ca-9e8e-a2853699f8b6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.446635] env[63088]: DEBUG oslo_vmware.api [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] Task: {'id': task-1284794, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 584.699589] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f5f50761-0430-407c-84cc-7b2c9cf8683b tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 584.920249] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance ca92ddb4-cac8-450a-9b7e-d09ad082354a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 584.931377] env[63088]: DEBUG oslo_vmware.api [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] Task: {'id': task-1284794, 'name': PowerOffVM_Task, 'duration_secs': 0.116453} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 584.932165] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 584.932165] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 584.933100] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dac7062c-334f-418a-b12f-195f394fdd99 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.940457] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Unregistering the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 584.942378] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7e39f041-d8a8-4f98-a40c-97402a548b08 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.975848] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Unregistered the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 584.976550] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Deleting contents of the VM from datastore datastore2 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 584.976550] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] Deleting the datastore file [datastore2] e8cfbce4-cfb2-463e-9f38-78a058dab6e1 {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 584.976550] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-42e4419c-49dc-42bb-b3f7-a3d161060e65 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.985770] env[63088]: DEBUG oslo_vmware.api [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] Waiting for the task: (returnval){ [ 584.985770] env[63088]: value = "task-1284796" [ 584.985770] env[63088]: _type = "Task" [ 584.985770] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 584.997114] env[63088]: DEBUG oslo_vmware.api [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] Task: {'id': task-1284796, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 585.078665] env[63088]: DEBUG oslo_concurrency.lockutils [None req-442e0a41-8be2-46c5-9922-69b38e5ecc58 tempest-ServerAddressesNegativeTestJSON-62458283 tempest-ServerAddressesNegativeTestJSON-62458283-project-member] Acquiring lock "be5c7b15-c036-4a20-a21b-74bba67c9a2c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 585.079771] env[63088]: DEBUG oslo_concurrency.lockutils [None req-442e0a41-8be2-46c5-9922-69b38e5ecc58 tempest-ServerAddressesNegativeTestJSON-62458283 tempest-ServerAddressesNegativeTestJSON-62458283-project-member] Lock "be5c7b15-c036-4a20-a21b-74bba67c9a2c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 585.427052] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance d6c76851-7059-40a4-a39b-e625c99dfcd9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 585.499184] env[63088]: DEBUG oslo_vmware.api [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] Task: {'id': task-1284796, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.309278} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 585.499517] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 585.499699] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Deleted contents of the VM from datastore datastore2 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 585.499958] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 585.936516] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance 91a09343-c2d7-4f5d-ae0a-4a123af92fba has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 586.439543] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance f29f6a5e-8621-4802-a679-d5c9a8fb2461 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 586.539212] env[63088]: ERROR nova.compute.manager [None req-d1272b04-ee5c-4e9e-b0bf-6ca9d0629aae tempest-VolumesAssistedSnapshotsTest-1004656420 tempest-VolumesAssistedSnapshotsTest-1004656420-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e0dd550d-f2d8-47df-9706-a8988ec41e69, please check neutron logs for more information. [ 586.539212] env[63088]: ERROR nova.compute.manager Traceback (most recent call last): [ 586.539212] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 586.539212] env[63088]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 586.539212] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 586.539212] env[63088]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 586.539212] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 586.539212] env[63088]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 586.539212] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 586.539212] env[63088]: ERROR nova.compute.manager self.force_reraise() [ 586.539212] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 586.539212] env[63088]: ERROR nova.compute.manager raise self.value [ 586.539212] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 586.539212] env[63088]: ERROR nova.compute.manager updated_port = self._update_port( [ 586.539212] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 586.539212] env[63088]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 586.539674] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 586.539674] env[63088]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 586.539674] env[63088]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e0dd550d-f2d8-47df-9706-a8988ec41e69, please check neutron logs for more information. [ 586.539674] env[63088]: ERROR nova.compute.manager [ 586.539674] env[63088]: Traceback (most recent call last): [ 586.539674] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 586.539674] env[63088]: listener.cb(fileno) [ 586.539674] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 586.539674] env[63088]: result = function(*args, **kwargs) [ 586.539674] env[63088]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 586.539674] env[63088]: return func(*args, **kwargs) [ 586.539674] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 586.539674] env[63088]: raise e [ 586.539674] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 586.539674] env[63088]: nwinfo = self.network_api.allocate_for_instance( [ 586.539674] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 586.539674] env[63088]: created_port_ids = self._update_ports_for_instance( [ 586.539674] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 586.539674] env[63088]: with excutils.save_and_reraise_exception(): [ 586.539674] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 586.539674] env[63088]: self.force_reraise() [ 586.539674] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 586.539674] env[63088]: raise self.value [ 586.539674] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 586.539674] env[63088]: updated_port = self._update_port( [ 586.539674] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 586.539674] env[63088]: _ensure_no_port_binding_failure(port) [ 586.539674] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 586.539674] env[63088]: raise exception.PortBindingFailed(port_id=port['id']) [ 586.540537] env[63088]: nova.exception.PortBindingFailed: Binding failed for port e0dd550d-f2d8-47df-9706-a8988ec41e69, please check neutron logs for more information. [ 586.540537] env[63088]: Removing descriptor: 17 [ 586.540537] env[63088]: ERROR nova.compute.manager [None req-d1272b04-ee5c-4e9e-b0bf-6ca9d0629aae tempest-VolumesAssistedSnapshotsTest-1004656420 tempest-VolumesAssistedSnapshotsTest-1004656420-project-member] [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e0dd550d-f2d8-47df-9706-a8988ec41e69, please check neutron logs for more information. [ 586.540537] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] Traceback (most recent call last): [ 586.540537] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 586.540537] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] yield resources [ 586.540537] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 586.540537] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] self.driver.spawn(context, instance, image_meta, [ 586.540537] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 586.540537] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 586.540537] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 586.540537] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] vm_ref = self.build_virtual_machine(instance, [ 586.540972] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 586.540972] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] vif_infos = vmwarevif.get_vif_info(self._session, [ 586.540972] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 586.540972] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] for vif in network_info: [ 586.540972] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 586.540972] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] return self._sync_wrapper(fn, *args, **kwargs) [ 586.540972] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 586.540972] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] self.wait() [ 586.540972] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 586.540972] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] self[:] = self._gt.wait() [ 586.540972] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 586.540972] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] return self._exit_event.wait() [ 586.540972] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 586.541584] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] result = hub.switch() [ 586.541584] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 586.541584] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] return self.greenlet.switch() [ 586.541584] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 586.541584] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] result = function(*args, **kwargs) [ 586.541584] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 586.541584] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] return func(*args, **kwargs) [ 586.541584] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 586.541584] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] raise e [ 586.541584] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 586.541584] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] nwinfo = self.network_api.allocate_for_instance( [ 586.541584] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 586.541584] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] created_port_ids = self._update_ports_for_instance( [ 586.541967] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 586.541967] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] with excutils.save_and_reraise_exception(): [ 586.541967] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 586.541967] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] self.force_reraise() [ 586.541967] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 586.541967] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] raise self.value [ 586.541967] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 586.541967] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] updated_port = self._update_port( [ 586.541967] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 586.541967] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] _ensure_no_port_binding_failure(port) [ 586.541967] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 586.541967] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] raise exception.PortBindingFailed(port_id=port['id']) [ 586.542557] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] nova.exception.PortBindingFailed: Binding failed for port e0dd550d-f2d8-47df-9706-a8988ec41e69, please check neutron logs for more information. [ 586.542557] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] [ 586.542557] env[63088]: INFO nova.compute.manager [None req-d1272b04-ee5c-4e9e-b0bf-6ca9d0629aae tempest-VolumesAssistedSnapshotsTest-1004656420 tempest-VolumesAssistedSnapshotsTest-1004656420-project-member] [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] Terminating instance [ 586.545674] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d1272b04-ee5c-4e9e-b0bf-6ca9d0629aae tempest-VolumesAssistedSnapshotsTest-1004656420 tempest-VolumesAssistedSnapshotsTest-1004656420-project-member] Acquiring lock "refresh_cache-d2c07175-b635-4900-9dab-9767e5f6bce4" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 586.546023] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d1272b04-ee5c-4e9e-b0bf-6ca9d0629aae tempest-VolumesAssistedSnapshotsTest-1004656420 tempest-VolumesAssistedSnapshotsTest-1004656420-project-member] Acquired lock "refresh_cache-d2c07175-b635-4900-9dab-9767e5f6bce4" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 586.546272] env[63088]: DEBUG nova.network.neutron [None req-d1272b04-ee5c-4e9e-b0bf-6ca9d0629aae tempest-VolumesAssistedSnapshotsTest-1004656420 tempest-VolumesAssistedSnapshotsTest-1004656420-project-member] [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 586.557997] env[63088]: DEBUG nova.virt.hardware [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 586.558250] env[63088]: DEBUG nova.virt.hardware [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 586.558401] env[63088]: DEBUG nova.virt.hardware [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 586.558571] env[63088]: DEBUG nova.virt.hardware [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 586.558706] env[63088]: DEBUG nova.virt.hardware [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 586.558845] env[63088]: DEBUG nova.virt.hardware [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 586.559215] env[63088]: DEBUG nova.virt.hardware [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 586.559418] env[63088]: DEBUG nova.virt.hardware [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 586.559703] env[63088]: DEBUG nova.virt.hardware [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 586.561823] env[63088]: DEBUG nova.virt.hardware [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 586.561823] env[63088]: DEBUG nova.virt.hardware [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 586.562265] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b67172f1-f334-4cdb-afe0-3d2c0e5cdd52 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.571897] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-516ccc87-68e1-4a28-b43e-dde2e82eef38 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.588240] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Instance VIF info [] {{(pid=63088) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 586.595061] env[63088]: DEBUG oslo.service.loopingcall [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 586.595194] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Creating VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 586.595681] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1c48c92d-9799-4efd-8809-9e7e7ca7730d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.618553] env[63088]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 586.618553] env[63088]: value = "task-1284797" [ 586.618553] env[63088]: _type = "Task" [ 586.618553] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 586.627120] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1284797, 'name': CreateVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 586.640338] env[63088]: DEBUG nova.compute.manager [req-5413157b-be32-406a-a50a-1f923c6e2a6c req-292dd787-063a-4fc3-bda8-b57191ef7bc8 service nova] [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] Received event network-changed-e0dd550d-f2d8-47df-9706-a8988ec41e69 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 586.640533] env[63088]: DEBUG nova.compute.manager [req-5413157b-be32-406a-a50a-1f923c6e2a6c req-292dd787-063a-4fc3-bda8-b57191ef7bc8 service nova] [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] Refreshing instance network info cache due to event network-changed-e0dd550d-f2d8-47df-9706-a8988ec41e69. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 586.640750] env[63088]: DEBUG oslo_concurrency.lockutils [req-5413157b-be32-406a-a50a-1f923c6e2a6c req-292dd787-063a-4fc3-bda8-b57191ef7bc8 service nova] Acquiring lock "refresh_cache-d2c07175-b635-4900-9dab-9767e5f6bce4" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 586.946023] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance 498c8e1e-18c6-4e09-a90d-106600bed2cd has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 587.074591] env[63088]: DEBUG nova.network.neutron [None req-d1272b04-ee5c-4e9e-b0bf-6ca9d0629aae tempest-VolumesAssistedSnapshotsTest-1004656420 tempest-VolumesAssistedSnapshotsTest-1004656420-project-member] [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 587.140120] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1284797, 'name': CreateVM_Task, 'duration_secs': 0.29387} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 587.140120] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Created VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 587.140120] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 587.140120] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 587.140120] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] Acquired external semaphore "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 587.140120] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-61509c1b-4e38-4ad2-adc3-98a8e0ce7d91 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.148813] env[63088]: DEBUG oslo_vmware.api [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] Waiting for the task: (returnval){ [ 587.148813] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]528cb192-bbc2-4271-2cc8-1e6fab03d737" [ 587.148813] env[63088]: _type = "Task" [ 587.148813] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 587.160118] env[63088]: DEBUG oslo_vmware.api [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]528cb192-bbc2-4271-2cc8-1e6fab03d737, 'name': SearchDatastore_Task, 'duration_secs': 0.010976} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 587.160410] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 587.160629] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Processing image 1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 587.160882] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 587.160996] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 587.161183] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 587.161431] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a39580c2-35e5-4941-a64a-3191f162cccc {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.169751] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 587.170791] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] Folder [datastore1] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 587.171562] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bd6981e0-2d57-49c5-a948-878f02f93160 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.179412] env[63088]: DEBUG oslo_vmware.api [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] Waiting for the task: (returnval){ [ 587.179412] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]521f106e-c31b-1aec-edfe-038df9c7e31b" [ 587.179412] env[63088]: _type = "Task" [ 587.179412] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 587.190257] env[63088]: DEBUG oslo_vmware.api [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]521f106e-c31b-1aec-edfe-038df9c7e31b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 587.352736] env[63088]: DEBUG nova.network.neutron [None req-d1272b04-ee5c-4e9e-b0bf-6ca9d0629aae tempest-VolumesAssistedSnapshotsTest-1004656420 tempest-VolumesAssistedSnapshotsTest-1004656420-project-member] [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 587.451712] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance 421e0f76-024d-4793-8f00-98731bb6b1af has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 587.701456] env[63088]: DEBUG oslo_vmware.api [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]521f106e-c31b-1aec-edfe-038df9c7e31b, 'name': SearchDatastore_Task, 'duration_secs': 0.023451} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 587.702948] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-07ed9e68-72d2-480b-8afd-9870de5dcc9a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.709132] env[63088]: DEBUG oslo_vmware.api [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] Waiting for the task: (returnval){ [ 587.709132] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]527918bc-25a5-54e7-2e85-a40206ddf202" [ 587.709132] env[63088]: _type = "Task" [ 587.709132] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 587.721346] env[63088]: DEBUG oslo_vmware.api [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]527918bc-25a5-54e7-2e85-a40206ddf202, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 587.858458] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d1272b04-ee5c-4e9e-b0bf-6ca9d0629aae tempest-VolumesAssistedSnapshotsTest-1004656420 tempest-VolumesAssistedSnapshotsTest-1004656420-project-member] Releasing lock "refresh_cache-d2c07175-b635-4900-9dab-9767e5f6bce4" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 587.858792] env[63088]: DEBUG nova.compute.manager [None req-d1272b04-ee5c-4e9e-b0bf-6ca9d0629aae tempest-VolumesAssistedSnapshotsTest-1004656420 tempest-VolumesAssistedSnapshotsTest-1004656420-project-member] [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 587.859017] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-d1272b04-ee5c-4e9e-b0bf-6ca9d0629aae tempest-VolumesAssistedSnapshotsTest-1004656420 tempest-VolumesAssistedSnapshotsTest-1004656420-project-member] [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 587.862407] env[63088]: DEBUG oslo_concurrency.lockutils [req-5413157b-be32-406a-a50a-1f923c6e2a6c req-292dd787-063a-4fc3-bda8-b57191ef7bc8 service nova] Acquired lock "refresh_cache-d2c07175-b635-4900-9dab-9767e5f6bce4" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 587.862550] env[63088]: DEBUG nova.network.neutron [req-5413157b-be32-406a-a50a-1f923c6e2a6c req-292dd787-063a-4fc3-bda8-b57191ef7bc8 service nova] [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] Refreshing network info cache for port e0dd550d-f2d8-47df-9706-a8988ec41e69 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 587.865981] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9ff0c002-261d-426d-9e8d-f2172de8eecc {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.886099] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2beb993-626e-4cec-a473-1baad4fd672b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.920782] env[63088]: WARNING nova.virt.vmwareapi.vmops [None req-d1272b04-ee5c-4e9e-b0bf-6ca9d0629aae tempest-VolumesAssistedSnapshotsTest-1004656420 tempest-VolumesAssistedSnapshotsTest-1004656420-project-member] [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d2c07175-b635-4900-9dab-9767e5f6bce4 could not be found. [ 587.921097] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-d1272b04-ee5c-4e9e-b0bf-6ca9d0629aae tempest-VolumesAssistedSnapshotsTest-1004656420 tempest-VolumesAssistedSnapshotsTest-1004656420-project-member] [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 587.925021] env[63088]: INFO nova.compute.manager [None req-d1272b04-ee5c-4e9e-b0bf-6ca9d0629aae tempest-VolumesAssistedSnapshotsTest-1004656420 tempest-VolumesAssistedSnapshotsTest-1004656420-project-member] [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] Took 0.06 seconds to destroy the instance on the hypervisor. [ 587.925021] env[63088]: DEBUG oslo.service.loopingcall [None req-d1272b04-ee5c-4e9e-b0bf-6ca9d0629aae tempest-VolumesAssistedSnapshotsTest-1004656420 tempest-VolumesAssistedSnapshotsTest-1004656420-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 587.925021] env[63088]: DEBUG nova.compute.manager [-] [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 587.925021] env[63088]: DEBUG nova.network.neutron [-] [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 587.956158] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance 5b778247-1434-447e-8ff1-7359bb9e8ea2 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 588.042114] env[63088]: DEBUG nova.network.neutron [-] [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 588.230663] env[63088]: DEBUG oslo_vmware.api [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]527918bc-25a5-54e7-2e85-a40206ddf202, 'name': SearchDatastore_Task, 'duration_secs': 0.008913} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 588.230663] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 588.230663] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] Copying Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] e8cfbce4-cfb2-463e-9f38-78a058dab6e1/e8cfbce4-cfb2-463e-9f38-78a058dab6e1.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 588.230663] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2573c638-cb59-467a-b8fe-6f801fc9c96f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.236518] env[63088]: DEBUG oslo_vmware.api [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] Waiting for the task: (returnval){ [ 588.236518] env[63088]: value = "task-1284798" [ 588.236518] env[63088]: _type = "Task" [ 588.236518] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 588.245308] env[63088]: DEBUG oslo_vmware.api [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] Task: {'id': task-1284798, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 588.433793] env[63088]: DEBUG nova.network.neutron [req-5413157b-be32-406a-a50a-1f923c6e2a6c req-292dd787-063a-4fc3-bda8-b57191ef7bc8 service nova] [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 588.466836] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance 966fe303-09b9-449f-8bda-4f6cda135f7a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 588.544705] env[63088]: DEBUG nova.network.neutron [-] [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 588.750786] env[63088]: DEBUG oslo_vmware.api [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] Task: {'id': task-1284798, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.493694} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 588.751231] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] Copied Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] e8cfbce4-cfb2-463e-9f38-78a058dab6e1/e8cfbce4-cfb2-463e-9f38-78a058dab6e1.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 588.751486] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Extending root virtual disk to 1048576 {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 588.751769] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c4585761-91c3-4a1c-9b61-32b470bb0836 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.763949] env[63088]: DEBUG oslo_vmware.api [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] Waiting for the task: (returnval){ [ 588.763949] env[63088]: value = "task-1284799" [ 588.763949] env[63088]: _type = "Task" [ 588.763949] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 588.764866] env[63088]: DEBUG nova.network.neutron [req-5413157b-be32-406a-a50a-1f923c6e2a6c req-292dd787-063a-4fc3-bda8-b57191ef7bc8 service nova] [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 588.776649] env[63088]: DEBUG oslo_vmware.api [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] Task: {'id': task-1284799, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 588.972438] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance 62bd4013-f007-4f69-b037-14bedf1e63cf has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 589.019842] env[63088]: DEBUG nova.compute.manager [req-1472480f-9498-4b2e-9f29-220e6630887e req-d3b28436-9b4f-4e83-b64a-4c666b74569c service nova] [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] Received event network-vif-deleted-e0dd550d-f2d8-47df-9706-a8988ec41e69 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 589.050613] env[63088]: INFO nova.compute.manager [-] [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] Took 1.13 seconds to deallocate network for instance. [ 589.053605] env[63088]: DEBUG nova.compute.claims [None req-d1272b04-ee5c-4e9e-b0bf-6ca9d0629aae tempest-VolumesAssistedSnapshotsTest-1004656420 tempest-VolumesAssistedSnapshotsTest-1004656420-project-member] [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] Aborting claim: {{(pid=63088) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 589.053743] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d1272b04-ee5c-4e9e-b0bf-6ca9d0629aae tempest-VolumesAssistedSnapshotsTest-1004656420 tempest-VolumesAssistedSnapshotsTest-1004656420-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 589.273828] env[63088]: DEBUG oslo_concurrency.lockutils [req-5413157b-be32-406a-a50a-1f923c6e2a6c req-292dd787-063a-4fc3-bda8-b57191ef7bc8 service nova] Releasing lock "refresh_cache-d2c07175-b635-4900-9dab-9767e5f6bce4" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 589.282123] env[63088]: DEBUG oslo_vmware.api [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] Task: {'id': task-1284799, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066218} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 589.282398] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Extended root virtual disk {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 589.284603] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d237cad8-fefb-41d4-98df-91eeff2046e4 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.312195] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Reconfiguring VM instance instance-00000005 to attach disk [datastore1] e8cfbce4-cfb2-463e-9f38-78a058dab6e1/e8cfbce4-cfb2-463e-9f38-78a058dab6e1.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 589.312511] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-00e50acf-db62-4281-be08-97dbbdfed220 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.335432] env[63088]: DEBUG oslo_vmware.api [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] Waiting for the task: (returnval){ [ 589.335432] env[63088]: value = "task-1284800" [ 589.335432] env[63088]: _type = "Task" [ 589.335432] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 589.346648] env[63088]: DEBUG oslo_vmware.api [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] Task: {'id': task-1284800, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 589.479970] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 589.480314] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Total usable vcpus: 48, total allocated vcpus: 6 {{(pid=63088) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 589.480314] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1664MB phys_disk=200GB used_disk=6GB total_vcpus=48 used_vcpus=6 pci_stats=[] {{(pid=63088) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 589.849256] env[63088]: DEBUG oslo_vmware.api [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] Task: {'id': task-1284800, 'name': ReconfigVM_Task, 'duration_secs': 0.333886} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 589.850592] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Reconfigured VM instance instance-00000005 to attach disk [datastore1] e8cfbce4-cfb2-463e-9f38-78a058dab6e1/e8cfbce4-cfb2-463e-9f38-78a058dab6e1.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 589.852074] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c86a366a-7c55-4c7c-9a82-9d69f8c5cba3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.854780] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c6878ab3-1216-435b-b7ff-23d8bbdc7851 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.863613] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b58c8f50-e4db-445b-80d7-fffdce57e465 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.866429] env[63088]: DEBUG oslo_vmware.api [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] Waiting for the task: (returnval){ [ 589.866429] env[63088]: value = "task-1284801" [ 589.866429] env[63088]: _type = "Task" [ 589.866429] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 589.896356] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-087fbe53-ff0a-4e36-b8cd-20e99e5eeea5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.902517] env[63088]: DEBUG oslo_vmware.api [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] Task: {'id': task-1284801, 'name': Rename_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 589.907420] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc9acb5f-d923-4400-94d2-5132e791f774 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.921389] env[63088]: DEBUG nova.compute.provider_tree [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 590.377374] env[63088]: DEBUG oslo_vmware.api [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] Task: {'id': task-1284801, 'name': Rename_Task, 'duration_secs': 0.146757} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 590.377559] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 590.377811] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2f5f609e-8f99-4d9b-8486-bc1cc8c2cd10 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.385655] env[63088]: DEBUG oslo_vmware.api [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] Waiting for the task: (returnval){ [ 590.385655] env[63088]: value = "task-1284802" [ 590.385655] env[63088]: _type = "Task" [ 590.385655] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 590.403460] env[63088]: DEBUG oslo_vmware.api [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] Task: {'id': task-1284802, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 590.425270] env[63088]: DEBUG nova.scheduler.client.report [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 590.901693] env[63088]: DEBUG oslo_vmware.api [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] Task: {'id': task-1284802, 'name': PowerOnVM_Task, 'duration_secs': 0.472868} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 590.902135] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 590.902180] env[63088]: DEBUG nova.compute.manager [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 590.903275] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-076a48a4-2a14-4a70-b71f-3b9d55dee28d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.933711] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63088) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 590.933711] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 8.593s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 590.933883] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6568dcbe-ebf7-46f4-8cbf-a6cec6f2c264 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.361s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 591.092215] env[63088]: DEBUG oslo_concurrency.lockutils [None req-32cec7a8-18a7-43df-b82f-d466be0dcd4a tempest-AttachInterfacesUnderV243Test-1205880501 tempest-AttachInterfacesUnderV243Test-1205880501-project-member] Acquiring lock "6137d44e-9945-479d-9cf3-4a36bfa0e32c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 591.092215] env[63088]: DEBUG oslo_concurrency.lockutils [None req-32cec7a8-18a7-43df-b82f-d466be0dcd4a tempest-AttachInterfacesUnderV243Test-1205880501 tempest-AttachInterfacesUnderV243Test-1205880501-project-member] Lock "6137d44e-9945-479d-9cf3-4a36bfa0e32c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 591.424869] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 591.743093] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bb8843aa-5c19-4289-a41c-17cbfcb47a00 tempest-ServerActionsTestOtherA-1614359098 tempest-ServerActionsTestOtherA-1614359098-project-member] Acquiring lock "90c1832a-47c5-4f09-9921-21e15cb6fbb6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 591.744938] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bb8843aa-5c19-4289-a41c-17cbfcb47a00 tempest-ServerActionsTestOtherA-1614359098 tempest-ServerActionsTestOtherA-1614359098-project-member] Lock "90c1832a-47c5-4f09-9921-21e15cb6fbb6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 591.839430] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2512139-cdd6-421e-9513-50ec86a35598 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.855803] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33982ce6-0b2c-4eb2-b97c-28765eec6f4a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.899411] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f813a1c-fe07-4780-be60-0a0f60cafbcf {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.907867] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76a7f572-67fa-4bb6-ae9f-75afa3b75ff0 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.924920] env[63088]: DEBUG nova.compute.provider_tree [None req-6568dcbe-ebf7-46f4-8cbf-a6cec6f2c264 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 592.237207] env[63088]: DEBUG nova.compute.manager [None req-4beda469-9ef0-4af6-8c30-aca2d1fcaf98 tempest-ServerDiagnosticsV248Test-626672433 tempest-ServerDiagnosticsV248Test-626672433-project-admin] [instance: c4777721-3f65-455f-9973-c1ed0732de34] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 592.238915] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da09964f-acb4-4f9d-8d8e-451b1310c52e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.248044] env[63088]: INFO nova.compute.manager [None req-4beda469-9ef0-4af6-8c30-aca2d1fcaf98 tempest-ServerDiagnosticsV248Test-626672433 tempest-ServerDiagnosticsV248Test-626672433-project-admin] [instance: c4777721-3f65-455f-9973-c1ed0732de34] Retrieving diagnostics [ 592.248441] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7929722-e37c-4587-af21-fed2b2cf7682 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.428018] env[63088]: DEBUG nova.scheduler.client.report [None req-6568dcbe-ebf7-46f4-8cbf-a6cec6f2c264 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 592.614119] env[63088]: DEBUG oslo_concurrency.lockutils [None req-18116a16-f5bf-4db6-aaca-b64d4351d5d4 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Acquiring lock "0f515940-fa10-4eb8-970d-ead4b996021c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 592.614360] env[63088]: DEBUG oslo_concurrency.lockutils [None req-18116a16-f5bf-4db6-aaca-b64d4351d5d4 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Lock "0f515940-fa10-4eb8-970d-ead4b996021c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 592.938518] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6568dcbe-ebf7-46f4-8cbf-a6cec6f2c264 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.002s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 592.938518] env[63088]: ERROR nova.compute.manager [None req-6568dcbe-ebf7-46f4-8cbf-a6cec6f2c264 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] [instance: 4d5293a0-6945-4731-afb6-996a142c8447] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 0c9febcc-5137-47a2-a0b9-df090f78c2e7, please check neutron logs for more information. [ 592.938518] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] Traceback (most recent call last): [ 592.938518] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 592.938518] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] self.driver.spawn(context, instance, image_meta, [ 592.938518] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 592.938518] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] self._vmops.spawn(context, instance, image_meta, injected_files, [ 592.938518] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 592.938518] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] vm_ref = self.build_virtual_machine(instance, [ 592.939260] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 592.939260] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] vif_infos = vmwarevif.get_vif_info(self._session, [ 592.939260] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 592.939260] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] for vif in network_info: [ 592.939260] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 592.939260] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] return self._sync_wrapper(fn, *args, **kwargs) [ 592.939260] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 592.939260] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] self.wait() [ 592.939260] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 592.939260] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] self[:] = self._gt.wait() [ 592.939260] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 592.939260] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] return self._exit_event.wait() [ 592.939260] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 592.939621] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] result = hub.switch() [ 592.939621] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 592.939621] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] return self.greenlet.switch() [ 592.939621] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 592.939621] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] result = function(*args, **kwargs) [ 592.939621] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 592.939621] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] return func(*args, **kwargs) [ 592.939621] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 592.939621] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] raise e [ 592.939621] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 592.939621] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] nwinfo = self.network_api.allocate_for_instance( [ 592.939621] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 592.939621] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] created_port_ids = self._update_ports_for_instance( [ 592.940014] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 592.940014] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] with excutils.save_and_reraise_exception(): [ 592.940014] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 592.940014] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] self.force_reraise() [ 592.940014] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 592.940014] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] raise self.value [ 592.940014] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 592.940014] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] updated_port = self._update_port( [ 592.940014] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 592.940014] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] _ensure_no_port_binding_failure(port) [ 592.940014] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 592.940014] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] raise exception.PortBindingFailed(port_id=port['id']) [ 592.940368] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] nova.exception.PortBindingFailed: Binding failed for port 0c9febcc-5137-47a2-a0b9-df090f78c2e7, please check neutron logs for more information. [ 592.940368] env[63088]: ERROR nova.compute.manager [instance: 4d5293a0-6945-4731-afb6-996a142c8447] [ 592.940368] env[63088]: DEBUG nova.compute.utils [None req-6568dcbe-ebf7-46f4-8cbf-a6cec6f2c264 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] [instance: 4d5293a0-6945-4731-afb6-996a142c8447] Binding failed for port 0c9febcc-5137-47a2-a0b9-df090f78c2e7, please check neutron logs for more information. {{(pid=63088) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 592.940738] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bc610745-9520-42a6-9fd1-7e4c9c63b77a tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.992s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 592.944637] env[63088]: DEBUG nova.compute.manager [None req-6568dcbe-ebf7-46f4-8cbf-a6cec6f2c264 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] [instance: 4d5293a0-6945-4731-afb6-996a142c8447] Build of instance 4d5293a0-6945-4731-afb6-996a142c8447 was re-scheduled: Binding failed for port 0c9febcc-5137-47a2-a0b9-df090f78c2e7, please check neutron logs for more information. {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 592.945423] env[63088]: DEBUG nova.compute.manager [None req-6568dcbe-ebf7-46f4-8cbf-a6cec6f2c264 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] [instance: 4d5293a0-6945-4731-afb6-996a142c8447] Unplugging VIFs for instance {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 592.945423] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6568dcbe-ebf7-46f4-8cbf-a6cec6f2c264 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] Acquiring lock "refresh_cache-4d5293a0-6945-4731-afb6-996a142c8447" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 592.945532] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6568dcbe-ebf7-46f4-8cbf-a6cec6f2c264 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] Acquired lock "refresh_cache-4d5293a0-6945-4731-afb6-996a142c8447" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 592.946221] env[63088]: DEBUG nova.network.neutron [None req-6568dcbe-ebf7-46f4-8cbf-a6cec6f2c264 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] [instance: 4d5293a0-6945-4731-afb6-996a142c8447] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 593.515361] env[63088]: DEBUG nova.network.neutron [None req-6568dcbe-ebf7-46f4-8cbf-a6cec6f2c264 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] [instance: 4d5293a0-6945-4731-afb6-996a142c8447] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 593.534389] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0e2f1d1d-62ce-4a58-953f-25113ba15c0f tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Acquiring lock "e8cfbce4-cfb2-463e-9f38-78a058dab6e1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 593.534389] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0e2f1d1d-62ce-4a58-953f-25113ba15c0f tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Lock "e8cfbce4-cfb2-463e-9f38-78a058dab6e1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 593.534389] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0e2f1d1d-62ce-4a58-953f-25113ba15c0f tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Acquiring lock "e8cfbce4-cfb2-463e-9f38-78a058dab6e1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 593.534389] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0e2f1d1d-62ce-4a58-953f-25113ba15c0f tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Lock "e8cfbce4-cfb2-463e-9f38-78a058dab6e1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 593.536657] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0e2f1d1d-62ce-4a58-953f-25113ba15c0f tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Lock "e8cfbce4-cfb2-463e-9f38-78a058dab6e1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 593.540322] env[63088]: INFO nova.compute.manager [None req-0e2f1d1d-62ce-4a58-953f-25113ba15c0f tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Terminating instance [ 593.543066] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0e2f1d1d-62ce-4a58-953f-25113ba15c0f tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Acquiring lock "refresh_cache-e8cfbce4-cfb2-463e-9f38-78a058dab6e1" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 593.543066] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0e2f1d1d-62ce-4a58-953f-25113ba15c0f tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Acquired lock "refresh_cache-e8cfbce4-cfb2-463e-9f38-78a058dab6e1" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 593.543066] env[63088]: DEBUG nova.network.neutron [None req-0e2f1d1d-62ce-4a58-953f-25113ba15c0f tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 593.682947] env[63088]: DEBUG nova.network.neutron [None req-6568dcbe-ebf7-46f4-8cbf-a6cec6f2c264 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] [instance: 4d5293a0-6945-4731-afb6-996a142c8447] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 593.885786] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-958b31ee-261e-4337-8f57-054ab3fe3f05 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.898894] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88e273cd-cff2-4c95-aaa3-521d052500a6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.959078] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4af384bd-cb67-4835-85fc-f744ba79c020 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.973773] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa889e0d-a889-44c0-a228-612204d57427 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.995372] env[63088]: DEBUG nova.compute.provider_tree [None req-bc610745-9520-42a6-9fd1-7e4c9c63b77a tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 594.006137] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f237740d-6eaf-4423-b770-dd9775fe6fd5 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Acquiring lock "d9a4d198-4aee-4882-bee8-f6cde14745a2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 594.006137] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f237740d-6eaf-4423-b770-dd9775fe6fd5 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Lock "d9a4d198-4aee-4882-bee8-f6cde14745a2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 594.086132] env[63088]: DEBUG nova.network.neutron [None req-0e2f1d1d-62ce-4a58-953f-25113ba15c0f tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 594.185314] env[63088]: DEBUG nova.network.neutron [None req-0e2f1d1d-62ce-4a58-953f-25113ba15c0f tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 594.187914] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6568dcbe-ebf7-46f4-8cbf-a6cec6f2c264 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] Releasing lock "refresh_cache-4d5293a0-6945-4731-afb6-996a142c8447" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 594.187914] env[63088]: DEBUG nova.compute.manager [None req-6568dcbe-ebf7-46f4-8cbf-a6cec6f2c264 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 594.187914] env[63088]: DEBUG nova.compute.manager [None req-6568dcbe-ebf7-46f4-8cbf-a6cec6f2c264 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] [instance: 4d5293a0-6945-4731-afb6-996a142c8447] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 594.187914] env[63088]: DEBUG nova.network.neutron [None req-6568dcbe-ebf7-46f4-8cbf-a6cec6f2c264 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] [instance: 4d5293a0-6945-4731-afb6-996a142c8447] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 594.214557] env[63088]: DEBUG nova.network.neutron [None req-6568dcbe-ebf7-46f4-8cbf-a6cec6f2c264 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] [instance: 4d5293a0-6945-4731-afb6-996a142c8447] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 594.469842] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a1a58940-45a1-49d0-85a7-c4e62f4fb6f2 tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Acquiring lock "c4777721-3f65-455f-9973-c1ed0732de34" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 594.471139] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a1a58940-45a1-49d0-85a7-c4e62f4fb6f2 tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Lock "c4777721-3f65-455f-9973-c1ed0732de34" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 594.471431] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a1a58940-45a1-49d0-85a7-c4e62f4fb6f2 tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Acquiring lock "c4777721-3f65-455f-9973-c1ed0732de34-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 594.472343] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a1a58940-45a1-49d0-85a7-c4e62f4fb6f2 tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Lock "c4777721-3f65-455f-9973-c1ed0732de34-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 594.472343] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a1a58940-45a1-49d0-85a7-c4e62f4fb6f2 tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Lock "c4777721-3f65-455f-9973-c1ed0732de34-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 594.479036] env[63088]: INFO nova.compute.manager [None req-a1a58940-45a1-49d0-85a7-c4e62f4fb6f2 tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] [instance: c4777721-3f65-455f-9973-c1ed0732de34] Terminating instance [ 594.480649] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a1a58940-45a1-49d0-85a7-c4e62f4fb6f2 tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Acquiring lock "refresh_cache-c4777721-3f65-455f-9973-c1ed0732de34" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 594.480836] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a1a58940-45a1-49d0-85a7-c4e62f4fb6f2 tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Acquired lock "refresh_cache-c4777721-3f65-455f-9973-c1ed0732de34" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 594.481029] env[63088]: DEBUG nova.network.neutron [None req-a1a58940-45a1-49d0-85a7-c4e62f4fb6f2 tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] [instance: c4777721-3f65-455f-9973-c1ed0732de34] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 594.500759] env[63088]: DEBUG nova.scheduler.client.report [None req-bc610745-9520-42a6-9fd1-7e4c9c63b77a tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 594.597768] env[63088]: DEBUG oslo_concurrency.lockutils [None req-393edcfb-0595-4190-8ce0-4c14157dbbff tempest-ServerDiagnosticsNegativeTest-1508062492 tempest-ServerDiagnosticsNegativeTest-1508062492-project-member] Acquiring lock "6eb3e767-ff7d-4760-a036-a55647afac0f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 594.598029] env[63088]: DEBUG oslo_concurrency.lockutils [None req-393edcfb-0595-4190-8ce0-4c14157dbbff tempest-ServerDiagnosticsNegativeTest-1508062492 tempest-ServerDiagnosticsNegativeTest-1508062492-project-member] Lock "6eb3e767-ff7d-4760-a036-a55647afac0f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 594.598237] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0b29042d-57c3-47a9-a926-f67d685a4740 tempest-TenantUsagesTestJSON-1054626739 tempest-TenantUsagesTestJSON-1054626739-project-member] Acquiring lock "02da88f0-7e3a-4eea-a88d-59a990c69f82" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 594.598421] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0b29042d-57c3-47a9-a926-f67d685a4740 tempest-TenantUsagesTestJSON-1054626739 tempest-TenantUsagesTestJSON-1054626739-project-member] Lock "02da88f0-7e3a-4eea-a88d-59a990c69f82" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 594.690618] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0e2f1d1d-62ce-4a58-953f-25113ba15c0f tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Releasing lock "refresh_cache-e8cfbce4-cfb2-463e-9f38-78a058dab6e1" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 594.691058] env[63088]: DEBUG nova.compute.manager [None req-0e2f1d1d-62ce-4a58-953f-25113ba15c0f tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 594.691256] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-0e2f1d1d-62ce-4a58-953f-25113ba15c0f tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 594.692144] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-304fd6b4-1945-4bc0-8cb5-2dfc56a223b4 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.703019] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e2f1d1d-62ce-4a58-953f-25113ba15c0f tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 594.703019] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-60e3a59c-4310-4285-b465-390d3ec659ba {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.711700] env[63088]: DEBUG oslo_vmware.api [None req-0e2f1d1d-62ce-4a58-953f-25113ba15c0f tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Waiting for the task: (returnval){ [ 594.711700] env[63088]: value = "task-1284803" [ 594.711700] env[63088]: _type = "Task" [ 594.711700] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 594.726527] env[63088]: DEBUG nova.network.neutron [None req-6568dcbe-ebf7-46f4-8cbf-a6cec6f2c264 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] [instance: 4d5293a0-6945-4731-afb6-996a142c8447] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 594.727766] env[63088]: DEBUG oslo_vmware.api [None req-0e2f1d1d-62ce-4a58-953f-25113ba15c0f tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Task: {'id': task-1284803, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 595.009914] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bc610745-9520-42a6-9fd1-7e4c9c63b77a tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.069s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 595.010853] env[63088]: ERROR nova.compute.manager [None req-bc610745-9520-42a6-9fd1-7e4c9c63b77a tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 010a5358-1268-49c2-b2b4-9525552d6fdc, please check neutron logs for more information. [ 595.010853] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] Traceback (most recent call last): [ 595.010853] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 595.010853] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] self.driver.spawn(context, instance, image_meta, [ 595.010853] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 595.010853] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] self._vmops.spawn(context, instance, image_meta, injected_files, [ 595.010853] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 595.010853] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] vm_ref = self.build_virtual_machine(instance, [ 595.010853] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 595.010853] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] vif_infos = vmwarevif.get_vif_info(self._session, [ 595.010853] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 595.011252] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] for vif in network_info: [ 595.011252] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 595.011252] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] return self._sync_wrapper(fn, *args, **kwargs) [ 595.011252] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 595.011252] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] self.wait() [ 595.011252] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 595.011252] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] self[:] = self._gt.wait() [ 595.011252] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 595.011252] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] return self._exit_event.wait() [ 595.011252] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 595.011252] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] result = hub.switch() [ 595.011252] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 595.011252] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] return self.greenlet.switch() [ 595.011731] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 595.011731] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] result = function(*args, **kwargs) [ 595.011731] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 595.011731] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] return func(*args, **kwargs) [ 595.011731] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 595.011731] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] raise e [ 595.011731] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 595.011731] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] nwinfo = self.network_api.allocate_for_instance( [ 595.011731] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 595.011731] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] created_port_ids = self._update_ports_for_instance( [ 595.011731] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 595.011731] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] with excutils.save_and_reraise_exception(): [ 595.011731] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 595.012250] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] self.force_reraise() [ 595.012250] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 595.012250] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] raise self.value [ 595.012250] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 595.012250] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] updated_port = self._update_port( [ 595.012250] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 595.012250] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] _ensure_no_port_binding_failure(port) [ 595.012250] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 595.012250] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] raise exception.PortBindingFailed(port_id=port['id']) [ 595.012250] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] nova.exception.PortBindingFailed: Binding failed for port 010a5358-1268-49c2-b2b4-9525552d6fdc, please check neutron logs for more information. [ 595.012250] env[63088]: ERROR nova.compute.manager [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] [ 595.013679] env[63088]: DEBUG nova.compute.utils [None req-bc610745-9520-42a6-9fd1-7e4c9c63b77a tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] Binding failed for port 010a5358-1268-49c2-b2b4-9525552d6fdc, please check neutron logs for more information. {{(pid=63088) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 595.013679] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6e35ce22-cf37-483a-8093-9574de222716 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.327s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 595.019055] env[63088]: DEBUG nova.compute.manager [None req-bc610745-9520-42a6-9fd1-7e4c9c63b77a tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] Build of instance 5c3f72e3-d4f4-4c34-92a3-075536b85c99 was re-scheduled: Binding failed for port 010a5358-1268-49c2-b2b4-9525552d6fdc, please check neutron logs for more information. {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 595.019687] env[63088]: DEBUG nova.compute.manager [None req-bc610745-9520-42a6-9fd1-7e4c9c63b77a tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] Unplugging VIFs for instance {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 595.019883] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bc610745-9520-42a6-9fd1-7e4c9c63b77a tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] Acquiring lock "refresh_cache-5c3f72e3-d4f4-4c34-92a3-075536b85c99" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 595.020021] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bc610745-9520-42a6-9fd1-7e4c9c63b77a tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] Acquired lock "refresh_cache-5c3f72e3-d4f4-4c34-92a3-075536b85c99" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 595.020695] env[63088]: DEBUG nova.network.neutron [None req-bc610745-9520-42a6-9fd1-7e4c9c63b77a tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 595.025846] env[63088]: DEBUG nova.network.neutron [None req-a1a58940-45a1-49d0-85a7-c4e62f4fb6f2 tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] [instance: c4777721-3f65-455f-9973-c1ed0732de34] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 595.222822] env[63088]: DEBUG oslo_vmware.api [None req-0e2f1d1d-62ce-4a58-953f-25113ba15c0f tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Task: {'id': task-1284803, 'name': PowerOffVM_Task, 'duration_secs': 0.131096} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 595.223128] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e2f1d1d-62ce-4a58-953f-25113ba15c0f tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 595.223306] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-0e2f1d1d-62ce-4a58-953f-25113ba15c0f tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Unregistering the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 595.223553] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-eaef8495-7f00-4de2-bb5b-b92727a839ab {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.229068] env[63088]: INFO nova.compute.manager [None req-6568dcbe-ebf7-46f4-8cbf-a6cec6f2c264 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] [instance: 4d5293a0-6945-4731-afb6-996a142c8447] Took 1.04 seconds to deallocate network for instance. [ 595.258235] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-0e2f1d1d-62ce-4a58-953f-25113ba15c0f tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Unregistered the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 595.258469] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-0e2f1d1d-62ce-4a58-953f-25113ba15c0f tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Deleting contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 595.258825] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e2f1d1d-62ce-4a58-953f-25113ba15c0f tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Deleting the datastore file [datastore1] e8cfbce4-cfb2-463e-9f38-78a058dab6e1 {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 595.258912] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b21354ef-9999-4998-8c80-c6fa406ce0db {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.268932] env[63088]: DEBUG oslo_vmware.api [None req-0e2f1d1d-62ce-4a58-953f-25113ba15c0f tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Waiting for the task: (returnval){ [ 595.268932] env[63088]: value = "task-1284805" [ 595.268932] env[63088]: _type = "Task" [ 595.268932] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 595.277788] env[63088]: DEBUG oslo_vmware.api [None req-0e2f1d1d-62ce-4a58-953f-25113ba15c0f tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Task: {'id': task-1284805, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 595.394618] env[63088]: DEBUG nova.network.neutron [None req-a1a58940-45a1-49d0-85a7-c4e62f4fb6f2 tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] [instance: c4777721-3f65-455f-9973-c1ed0732de34] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 595.565325] env[63088]: DEBUG nova.network.neutron [None req-bc610745-9520-42a6-9fd1-7e4c9c63b77a tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 595.694058] env[63088]: DEBUG oslo_concurrency.lockutils [None req-fcd5880e-5b8a-42fa-b638-13cd9836ac96 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Acquiring lock "c67b276f-07db-4cde-9f1c-efccc4d11c7d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 595.694292] env[63088]: DEBUG oslo_concurrency.lockutils [None req-fcd5880e-5b8a-42fa-b638-13cd9836ac96 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Lock "c67b276f-07db-4cde-9f1c-efccc4d11c7d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 595.787564] env[63088]: DEBUG oslo_vmware.api [None req-0e2f1d1d-62ce-4a58-953f-25113ba15c0f tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Task: {'id': task-1284805, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.11935} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 595.787899] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e2f1d1d-62ce-4a58-953f-25113ba15c0f tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 595.790167] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-0e2f1d1d-62ce-4a58-953f-25113ba15c0f tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Deleted contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 595.790167] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-0e2f1d1d-62ce-4a58-953f-25113ba15c0f tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 595.790167] env[63088]: INFO nova.compute.manager [None req-0e2f1d1d-62ce-4a58-953f-25113ba15c0f tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Took 1.10 seconds to destroy the instance on the hypervisor. [ 595.790167] env[63088]: DEBUG oslo.service.loopingcall [None req-0e2f1d1d-62ce-4a58-953f-25113ba15c0f tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 595.790167] env[63088]: DEBUG nova.compute.manager [-] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 595.790167] env[63088]: DEBUG nova.network.neutron [-] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 595.813100] env[63088]: DEBUG nova.network.neutron [-] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 595.833313] env[63088]: DEBUG nova.network.neutron [None req-bc610745-9520-42a6-9fd1-7e4c9c63b77a tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 595.899842] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a1a58940-45a1-49d0-85a7-c4e62f4fb6f2 tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Releasing lock "refresh_cache-c4777721-3f65-455f-9973-c1ed0732de34" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 595.899842] env[63088]: DEBUG nova.compute.manager [None req-a1a58940-45a1-49d0-85a7-c4e62f4fb6f2 tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] [instance: c4777721-3f65-455f-9973-c1ed0732de34] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 595.899842] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-a1a58940-45a1-49d0-85a7-c4e62f4fb6f2 tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] [instance: c4777721-3f65-455f-9973-c1ed0732de34] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 595.901259] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-969fbbab-a6cf-4609-ab5c-bde853624dc7 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.910889] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1a58940-45a1-49d0-85a7-c4e62f4fb6f2 tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] [instance: c4777721-3f65-455f-9973-c1ed0732de34] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 595.913900] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-27fd4201-1905-4721-89b2-880fb634b772 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.925543] env[63088]: DEBUG oslo_vmware.api [None req-a1a58940-45a1-49d0-85a7-c4e62f4fb6f2 tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Waiting for the task: (returnval){ [ 595.925543] env[63088]: value = "task-1284806" [ 595.925543] env[63088]: _type = "Task" [ 595.925543] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 595.935053] env[63088]: DEBUG oslo_vmware.api [None req-a1a58940-45a1-49d0-85a7-c4e62f4fb6f2 tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Task: {'id': task-1284806, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 596.016873] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6adc4f65-26ab-4e1b-a95b-e2789b3e51eb {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.026842] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a49ffeef-8e73-4062-92ad-c13717febd59 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.058431] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8a63520-251e-445d-8ffb-5eb920165c90 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.066351] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66ef24e8-aa62-4b72-bc4b-d234c5d01bfc {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.080719] env[63088]: DEBUG nova.compute.provider_tree [None req-6e35ce22-cf37-483a-8093-9574de222716 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 596.273568] env[63088]: INFO nova.scheduler.client.report [None req-6568dcbe-ebf7-46f4-8cbf-a6cec6f2c264 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] Deleted allocations for instance 4d5293a0-6945-4731-afb6-996a142c8447 [ 596.315889] env[63088]: DEBUG nova.network.neutron [-] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 596.336126] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bc610745-9520-42a6-9fd1-7e4c9c63b77a tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] Releasing lock "refresh_cache-5c3f72e3-d4f4-4c34-92a3-075536b85c99" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 596.336464] env[63088]: DEBUG nova.compute.manager [None req-bc610745-9520-42a6-9fd1-7e4c9c63b77a tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 596.336712] env[63088]: DEBUG nova.compute.manager [None req-bc610745-9520-42a6-9fd1-7e4c9c63b77a tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 596.336946] env[63088]: DEBUG nova.network.neutron [None req-bc610745-9520-42a6-9fd1-7e4c9c63b77a tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 596.431607] env[63088]: DEBUG nova.network.neutron [None req-bc610745-9520-42a6-9fd1-7e4c9c63b77a tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 596.437679] env[63088]: DEBUG oslo_vmware.api [None req-a1a58940-45a1-49d0-85a7-c4e62f4fb6f2 tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Task: {'id': task-1284806, 'name': PowerOffVM_Task, 'duration_secs': 0.11875} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 596.438463] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1a58940-45a1-49d0-85a7-c4e62f4fb6f2 tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] [instance: c4777721-3f65-455f-9973-c1ed0732de34] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 596.439257] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-a1a58940-45a1-49d0-85a7-c4e62f4fb6f2 tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] [instance: c4777721-3f65-455f-9973-c1ed0732de34] Unregistering the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 596.439257] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f18b6a2c-5858-4208-9a20-cb46d09d76b5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.466411] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-a1a58940-45a1-49d0-85a7-c4e62f4fb6f2 tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] [instance: c4777721-3f65-455f-9973-c1ed0732de34] Unregistered the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 596.466411] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-a1a58940-45a1-49d0-85a7-c4e62f4fb6f2 tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] [instance: c4777721-3f65-455f-9973-c1ed0732de34] Deleting contents of the VM from datastore datastore2 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 596.466411] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-a1a58940-45a1-49d0-85a7-c4e62f4fb6f2 tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Deleting the datastore file [datastore2] c4777721-3f65-455f-9973-c1ed0732de34 {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 596.467072] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ad6e894e-814b-4d6f-ac65-7e4119ed06b5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.481099] env[63088]: DEBUG oslo_vmware.api [None req-a1a58940-45a1-49d0-85a7-c4e62f4fb6f2 tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Waiting for the task: (returnval){ [ 596.481099] env[63088]: value = "task-1284808" [ 596.481099] env[63088]: _type = "Task" [ 596.481099] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 596.498252] env[63088]: DEBUG oslo_vmware.api [None req-a1a58940-45a1-49d0-85a7-c4e62f4fb6f2 tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Task: {'id': task-1284808, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 596.583830] env[63088]: DEBUG nova.scheduler.client.report [None req-6e35ce22-cf37-483a-8093-9574de222716 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 596.787628] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6568dcbe-ebf7-46f4-8cbf-a6cec6f2c264 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] Lock "4d5293a0-6945-4731-afb6-996a142c8447" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 47.615s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 596.819859] env[63088]: INFO nova.compute.manager [-] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Took 1.03 seconds to deallocate network for instance. [ 596.939415] env[63088]: DEBUG nova.network.neutron [None req-bc610745-9520-42a6-9fd1-7e4c9c63b77a tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 596.992994] env[63088]: DEBUG oslo_vmware.api [None req-a1a58940-45a1-49d0-85a7-c4e62f4fb6f2 tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Task: {'id': task-1284808, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.104365} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 596.993491] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-a1a58940-45a1-49d0-85a7-c4e62f4fb6f2 tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 596.993795] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-a1a58940-45a1-49d0-85a7-c4e62f4fb6f2 tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] [instance: c4777721-3f65-455f-9973-c1ed0732de34] Deleted contents of the VM from datastore datastore2 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 596.994086] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-a1a58940-45a1-49d0-85a7-c4e62f4fb6f2 tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] [instance: c4777721-3f65-455f-9973-c1ed0732de34] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 596.994391] env[63088]: INFO nova.compute.manager [None req-a1a58940-45a1-49d0-85a7-c4e62f4fb6f2 tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] [instance: c4777721-3f65-455f-9973-c1ed0732de34] Took 1.09 seconds to destroy the instance on the hypervisor. [ 596.994738] env[63088]: DEBUG oslo.service.loopingcall [None req-a1a58940-45a1-49d0-85a7-c4e62f4fb6f2 tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 596.995199] env[63088]: DEBUG nova.compute.manager [-] [instance: c4777721-3f65-455f-9973-c1ed0732de34] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 596.995372] env[63088]: DEBUG nova.network.neutron [-] [instance: c4777721-3f65-455f-9973-c1ed0732de34] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 597.033023] env[63088]: DEBUG nova.network.neutron [-] [instance: c4777721-3f65-455f-9973-c1ed0732de34] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 597.089174] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6e35ce22-cf37-483a-8093-9574de222716 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.076s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 597.089909] env[63088]: ERROR nova.compute.manager [None req-6e35ce22-cf37-483a-8093-9574de222716 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 27f3a8b8-19ea-48a7-8323-608f6af9414c, please check neutron logs for more information. [ 597.089909] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] Traceback (most recent call last): [ 597.089909] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 597.089909] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] self.driver.spawn(context, instance, image_meta, [ 597.089909] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 597.089909] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] self._vmops.spawn(context, instance, image_meta, injected_files, [ 597.089909] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 597.089909] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] vm_ref = self.build_virtual_machine(instance, [ 597.089909] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 597.089909] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] vif_infos = vmwarevif.get_vif_info(self._session, [ 597.089909] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 597.090473] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] for vif in network_info: [ 597.090473] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 597.090473] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] return self._sync_wrapper(fn, *args, **kwargs) [ 597.090473] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 597.090473] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] self.wait() [ 597.090473] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 597.090473] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] self[:] = self._gt.wait() [ 597.090473] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 597.090473] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] return self._exit_event.wait() [ 597.090473] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 597.090473] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] result = hub.switch() [ 597.090473] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 597.090473] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] return self.greenlet.switch() [ 597.091218] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 597.091218] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] result = function(*args, **kwargs) [ 597.091218] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 597.091218] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] return func(*args, **kwargs) [ 597.091218] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 597.091218] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] raise e [ 597.091218] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 597.091218] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] nwinfo = self.network_api.allocate_for_instance( [ 597.091218] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 597.091218] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] created_port_ids = self._update_ports_for_instance( [ 597.091218] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 597.091218] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] with excutils.save_and_reraise_exception(): [ 597.091218] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 597.091770] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] self.force_reraise() [ 597.091770] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 597.091770] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] raise self.value [ 597.091770] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 597.091770] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] updated_port = self._update_port( [ 597.091770] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 597.091770] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] _ensure_no_port_binding_failure(port) [ 597.091770] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 597.091770] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] raise exception.PortBindingFailed(port_id=port['id']) [ 597.091770] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] nova.exception.PortBindingFailed: Binding failed for port 27f3a8b8-19ea-48a7-8323-608f6af9414c, please check neutron logs for more information. [ 597.091770] env[63088]: ERROR nova.compute.manager [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] [ 597.092205] env[63088]: DEBUG nova.compute.utils [None req-6e35ce22-cf37-483a-8093-9574de222716 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] Binding failed for port 27f3a8b8-19ea-48a7-8323-608f6af9414c, please check neutron logs for more information. {{(pid=63088) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 597.092205] env[63088]: DEBUG oslo_concurrency.lockutils [None req-79832ead-48a0-4aad-a47e-8e38c2a687da tempest-ServersTestJSON-106593364 tempest-ServersTestJSON-106593364-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.199s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 597.093298] env[63088]: INFO nova.compute.claims [None req-79832ead-48a0-4aad-a47e-8e38c2a687da tempest-ServersTestJSON-106593364 tempest-ServersTestJSON-106593364-project-member] [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 597.096008] env[63088]: DEBUG nova.compute.manager [None req-6e35ce22-cf37-483a-8093-9574de222716 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] Build of instance 44e39b2b-2bac-4a6e-9534-d2774d094931 was re-scheduled: Binding failed for port 27f3a8b8-19ea-48a7-8323-608f6af9414c, please check neutron logs for more information. {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 597.097909] env[63088]: DEBUG nova.compute.manager [None req-6e35ce22-cf37-483a-8093-9574de222716 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] Unplugging VIFs for instance {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 597.097909] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6e35ce22-cf37-483a-8093-9574de222716 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] Acquiring lock "refresh_cache-44e39b2b-2bac-4a6e-9534-d2774d094931" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 597.097909] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6e35ce22-cf37-483a-8093-9574de222716 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] Acquired lock "refresh_cache-44e39b2b-2bac-4a6e-9534-d2774d094931" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 597.097909] env[63088]: DEBUG nova.network.neutron [None req-6e35ce22-cf37-483a-8093-9574de222716 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 597.292649] env[63088]: DEBUG nova.compute.manager [None req-b1b885bf-cef9-4808-8f16-707ab04499b3 tempest-ServersTestManualDisk-1859582357 tempest-ServersTestManualDisk-1859582357-project-member] [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 597.331613] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0e2f1d1d-62ce-4a58-953f-25113ba15c0f tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 597.442355] env[63088]: INFO nova.compute.manager [None req-bc610745-9520-42a6-9fd1-7e4c9c63b77a tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] [instance: 5c3f72e3-d4f4-4c34-92a3-075536b85c99] Took 1.11 seconds to deallocate network for instance. [ 597.533014] env[63088]: DEBUG nova.network.neutron [-] [instance: c4777721-3f65-455f-9973-c1ed0732de34] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 597.627148] env[63088]: DEBUG nova.network.neutron [None req-6e35ce22-cf37-483a-8093-9574de222716 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 597.820260] env[63088]: DEBUG nova.network.neutron [None req-6e35ce22-cf37-483a-8093-9574de222716 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 597.829930] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b1b885bf-cef9-4808-8f16-707ab04499b3 tempest-ServersTestManualDisk-1859582357 tempest-ServersTestManualDisk-1859582357-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 598.015108] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6c63332a-ec25-45cc-9207-42e0be237d32 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquiring lock "1219fe86-f218-4a2b-84ba-4a377829386d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 598.015361] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6c63332a-ec25-45cc-9207-42e0be237d32 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Lock "1219fe86-f218-4a2b-84ba-4a377829386d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 598.036135] env[63088]: INFO nova.compute.manager [-] [instance: c4777721-3f65-455f-9973-c1ed0732de34] Took 1.04 seconds to deallocate network for instance. [ 598.326357] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6e35ce22-cf37-483a-8093-9574de222716 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] Releasing lock "refresh_cache-44e39b2b-2bac-4a6e-9534-d2774d094931" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 598.326712] env[63088]: DEBUG nova.compute.manager [None req-6e35ce22-cf37-483a-8093-9574de222716 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 598.326894] env[63088]: DEBUG nova.compute.manager [None req-6e35ce22-cf37-483a-8093-9574de222716 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 598.327071] env[63088]: DEBUG nova.network.neutron [None req-6e35ce22-cf37-483a-8093-9574de222716 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 598.358663] env[63088]: DEBUG nova.network.neutron [None req-6e35ce22-cf37-483a-8093-9574de222716 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 598.496966] env[63088]: INFO nova.scheduler.client.report [None req-bc610745-9520-42a6-9fd1-7e4c9c63b77a tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] Deleted allocations for instance 5c3f72e3-d4f4-4c34-92a3-075536b85c99 [ 598.550827] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a1a58940-45a1-49d0-85a7-c4e62f4fb6f2 tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 598.625967] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0c6bc062-cb90-4d50-aabf-8c0503771f7e tempest-InstanceActionsNegativeTestJSON-1719444358 tempest-InstanceActionsNegativeTestJSON-1719444358-project-member] Acquiring lock "20611ed4-97b3-4d6d-bef7-c13e1235c401" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 598.626232] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0c6bc062-cb90-4d50-aabf-8c0503771f7e tempest-InstanceActionsNegativeTestJSON-1719444358 tempest-InstanceActionsNegativeTestJSON-1719444358-project-member] Lock "20611ed4-97b3-4d6d-bef7-c13e1235c401" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 598.686705] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09618944-5f6c-4146-bfbf-f5df5970f540 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.694903] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc414624-6be4-4f81-b1b9-87389a4b7462 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.735790] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5e1bc9c-f2a5-4a37-9c1f-106f617b5d49 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.749741] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-622746fa-19ef-412b-855c-2db835f97137 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.764684] env[63088]: DEBUG nova.compute.provider_tree [None req-79832ead-48a0-4aad-a47e-8e38c2a687da tempest-ServersTestJSON-106593364 tempest-ServersTestJSON-106593364-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 598.868837] env[63088]: DEBUG nova.network.neutron [None req-6e35ce22-cf37-483a-8093-9574de222716 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 599.012622] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bc610745-9520-42a6-9fd1-7e4c9c63b77a tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] Lock "5c3f72e3-d4f4-4c34-92a3-075536b85c99" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 51.371s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 599.271420] env[63088]: DEBUG nova.scheduler.client.report [None req-79832ead-48a0-4aad-a47e-8e38c2a687da tempest-ServersTestJSON-106593364 tempest-ServersTestJSON-106593364-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 599.378650] env[63088]: INFO nova.compute.manager [None req-6e35ce22-cf37-483a-8093-9574de222716 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] [instance: 44e39b2b-2bac-4a6e-9534-d2774d094931] Took 1.05 seconds to deallocate network for instance. [ 599.391618] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a2a0ac90-0e7a-45d1-a070-f7ee6be01d68 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquiring lock "eb42c667-b0ae-4807-8f69-88dd671aa7d7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 599.392183] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a2a0ac90-0e7a-45d1-a070-f7ee6be01d68 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Lock "eb42c667-b0ae-4807-8f69-88dd671aa7d7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 599.515999] env[63088]: DEBUG nova.compute.manager [None req-0564aade-729d-4cf1-a59d-513384386c54 tempest-AttachInterfacesV270Test-1783457293 tempest-AttachInterfacesV270Test-1783457293-project-member] [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 599.779572] env[63088]: DEBUG oslo_concurrency.lockutils [None req-79832ead-48a0-4aad-a47e-8e38c2a687da tempest-ServersTestJSON-106593364 tempest-ServersTestJSON-106593364-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.688s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 599.780607] env[63088]: DEBUG nova.compute.manager [None req-79832ead-48a0-4aad-a47e-8e38c2a687da tempest-ServersTestJSON-106593364 tempest-ServersTestJSON-106593364-project-member] [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 599.785382] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a9773ab2-f8b9-4102-8e41-5ee881a8637e tempest-FloatingIPsAssociationNegativeTestJSON-189613961 tempest-FloatingIPsAssociationNegativeTestJSON-189613961-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.091s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 599.787878] env[63088]: INFO nova.compute.claims [None req-a9773ab2-f8b9-4102-8e41-5ee881a8637e tempest-FloatingIPsAssociationNegativeTestJSON-189613961 tempest-FloatingIPsAssociationNegativeTestJSON-189613961-project-member] [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 600.050569] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0564aade-729d-4cf1-a59d-513384386c54 tempest-AttachInterfacesV270Test-1783457293 tempest-AttachInterfacesV270Test-1783457293-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 600.296121] env[63088]: DEBUG nova.compute.utils [None req-79832ead-48a0-4aad-a47e-8e38c2a687da tempest-ServersTestJSON-106593364 tempest-ServersTestJSON-106593364-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 600.297732] env[63088]: DEBUG nova.compute.manager [None req-79832ead-48a0-4aad-a47e-8e38c2a687da tempest-ServersTestJSON-106593364 tempest-ServersTestJSON-106593364-project-member] [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 600.297732] env[63088]: DEBUG nova.network.neutron [None req-79832ead-48a0-4aad-a47e-8e38c2a687da tempest-ServersTestJSON-106593364 tempest-ServersTestJSON-106593364-project-member] [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 600.353031] env[63088]: DEBUG nova.policy [None req-79832ead-48a0-4aad-a47e-8e38c2a687da tempest-ServersTestJSON-106593364 tempest-ServersTestJSON-106593364-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b85fd207607645b38180aeaf3a2389ea', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f8999492ec5043c5a81c8ba2408e6e9f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 600.426325] env[63088]: INFO nova.scheduler.client.report [None req-6e35ce22-cf37-483a-8093-9574de222716 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] Deleted allocations for instance 44e39b2b-2bac-4a6e-9534-d2774d094931 [ 600.804436] env[63088]: DEBUG nova.compute.manager [None req-79832ead-48a0-4aad-a47e-8e38c2a687da tempest-ServersTestJSON-106593364 tempest-ServersTestJSON-106593364-project-member] [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 600.942887] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6e35ce22-cf37-483a-8093-9574de222716 tempest-ServersAdminTestJSON-1429323702 tempest-ServersAdminTestJSON-1429323702-project-member] Lock "44e39b2b-2bac-4a6e-9534-d2774d094931" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 49.310s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 600.971772] env[63088]: DEBUG nova.network.neutron [None req-79832ead-48a0-4aad-a47e-8e38c2a687da tempest-ServersTestJSON-106593364 tempest-ServersTestJSON-106593364-project-member] [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] Successfully created port: 203eaeff-191a-4386-871b-58f5e69aa834 {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 601.149750] env[63088]: DEBUG oslo_concurrency.lockutils [None req-ae448e13-3c80-41d5-9b1f-e650107512d4 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquiring lock "2b43620d-94ac-4c56-87df-4066f5d24088" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 601.149990] env[63088]: DEBUG oslo_concurrency.lockutils [None req-ae448e13-3c80-41d5-9b1f-e650107512d4 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lock "2b43620d-94ac-4c56-87df-4066f5d24088" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 601.278658] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-176b8318-5261-468e-9b2e-1401a444a079 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.287082] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25d03d4e-251b-4ff9-a720-7ee66afd470e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.322163] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fac8b67-b2b6-47f1-8da1-a0e80e13b585 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.329207] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cccf714c-0779-43f2-bc86-41f5fc3b200b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.343083] env[63088]: DEBUG nova.compute.provider_tree [None req-a9773ab2-f8b9-4102-8e41-5ee881a8637e tempest-FloatingIPsAssociationNegativeTestJSON-189613961 tempest-FloatingIPsAssociationNegativeTestJSON-189613961-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 601.446073] env[63088]: DEBUG nova.compute.manager [None req-f81b332d-72eb-4149-a6a2-3c8a9c9bec53 tempest-ServersNegativeTestJSON-1008333314 tempest-ServersNegativeTestJSON-1008333314-project-member] [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 601.827952] env[63088]: DEBUG nova.compute.manager [None req-79832ead-48a0-4aad-a47e-8e38c2a687da tempest-ServersTestJSON-106593364 tempest-ServersTestJSON-106593364-project-member] [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 601.846052] env[63088]: DEBUG nova.scheduler.client.report [None req-a9773ab2-f8b9-4102-8e41-5ee881a8637e tempest-FloatingIPsAssociationNegativeTestJSON-189613961 tempest-FloatingIPsAssociationNegativeTestJSON-189613961-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 601.865121] env[63088]: DEBUG nova.virt.hardware [None req-79832ead-48a0-4aad-a47e-8e38c2a687da tempest-ServersTestJSON-106593364 tempest-ServersTestJSON-106593364-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 601.865121] env[63088]: DEBUG nova.virt.hardware [None req-79832ead-48a0-4aad-a47e-8e38c2a687da tempest-ServersTestJSON-106593364 tempest-ServersTestJSON-106593364-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 601.865121] env[63088]: DEBUG nova.virt.hardware [None req-79832ead-48a0-4aad-a47e-8e38c2a687da tempest-ServersTestJSON-106593364 tempest-ServersTestJSON-106593364-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 601.865121] env[63088]: DEBUG nova.virt.hardware [None req-79832ead-48a0-4aad-a47e-8e38c2a687da tempest-ServersTestJSON-106593364 tempest-ServersTestJSON-106593364-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 601.865719] env[63088]: DEBUG nova.virt.hardware [None req-79832ead-48a0-4aad-a47e-8e38c2a687da tempest-ServersTestJSON-106593364 tempest-ServersTestJSON-106593364-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 601.866186] env[63088]: DEBUG nova.virt.hardware [None req-79832ead-48a0-4aad-a47e-8e38c2a687da tempest-ServersTestJSON-106593364 tempest-ServersTestJSON-106593364-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 601.866549] env[63088]: DEBUG nova.virt.hardware [None req-79832ead-48a0-4aad-a47e-8e38c2a687da tempest-ServersTestJSON-106593364 tempest-ServersTestJSON-106593364-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 601.866779] env[63088]: DEBUG nova.virt.hardware [None req-79832ead-48a0-4aad-a47e-8e38c2a687da tempest-ServersTestJSON-106593364 tempest-ServersTestJSON-106593364-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 601.867044] env[63088]: DEBUG nova.virt.hardware [None req-79832ead-48a0-4aad-a47e-8e38c2a687da tempest-ServersTestJSON-106593364 tempest-ServersTestJSON-106593364-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 601.867320] env[63088]: DEBUG nova.virt.hardware [None req-79832ead-48a0-4aad-a47e-8e38c2a687da tempest-ServersTestJSON-106593364 tempest-ServersTestJSON-106593364-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 601.867595] env[63088]: DEBUG nova.virt.hardware [None req-79832ead-48a0-4aad-a47e-8e38c2a687da tempest-ServersTestJSON-106593364 tempest-ServersTestJSON-106593364-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 601.868779] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-751cf625-3ce5-4b0b-b066-d7749c4c2916 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.881647] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba7b4ac1-f260-41e4-b482-3ea78c98d855 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.973663] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f81b332d-72eb-4149-a6a2-3c8a9c9bec53 tempest-ServersNegativeTestJSON-1008333314 tempest-ServersNegativeTestJSON-1008333314-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 602.356718] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a9773ab2-f8b9-4102-8e41-5ee881a8637e tempest-FloatingIPsAssociationNegativeTestJSON-189613961 tempest-FloatingIPsAssociationNegativeTestJSON-189613961-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.571s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 602.357692] env[63088]: DEBUG nova.compute.manager [None req-a9773ab2-f8b9-4102-8e41-5ee881a8637e tempest-FloatingIPsAssociationNegativeTestJSON-189613961 tempest-FloatingIPsAssociationNegativeTestJSON-189613961-project-member] [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 602.364167] env[63088]: DEBUG oslo_concurrency.lockutils [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 22.060s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 602.364167] env[63088]: DEBUG nova.objects.instance [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63088) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 602.864311] env[63088]: DEBUG nova.compute.utils [None req-a9773ab2-f8b9-4102-8e41-5ee881a8637e tempest-FloatingIPsAssociationNegativeTestJSON-189613961 tempest-FloatingIPsAssociationNegativeTestJSON-189613961-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 602.866724] env[63088]: DEBUG nova.compute.manager [None req-a9773ab2-f8b9-4102-8e41-5ee881a8637e tempest-FloatingIPsAssociationNegativeTestJSON-189613961 tempest-FloatingIPsAssociationNegativeTestJSON-189613961-project-member] [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 602.866925] env[63088]: DEBUG nova.network.neutron [None req-a9773ab2-f8b9-4102-8e41-5ee881a8637e tempest-FloatingIPsAssociationNegativeTestJSON-189613961 tempest-FloatingIPsAssociationNegativeTestJSON-189613961-project-member] [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 602.985869] env[63088]: ERROR nova.compute.manager [None req-79832ead-48a0-4aad-a47e-8e38c2a687da tempest-ServersTestJSON-106593364 tempest-ServersTestJSON-106593364-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 203eaeff-191a-4386-871b-58f5e69aa834, please check neutron logs for more information. [ 602.985869] env[63088]: ERROR nova.compute.manager Traceback (most recent call last): [ 602.985869] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 602.985869] env[63088]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 602.985869] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 602.985869] env[63088]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 602.985869] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 602.985869] env[63088]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 602.985869] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 602.985869] env[63088]: ERROR nova.compute.manager self.force_reraise() [ 602.985869] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 602.985869] env[63088]: ERROR nova.compute.manager raise self.value [ 602.985869] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 602.985869] env[63088]: ERROR nova.compute.manager updated_port = self._update_port( [ 602.985869] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 602.985869] env[63088]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 602.986455] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 602.986455] env[63088]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 602.986455] env[63088]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 203eaeff-191a-4386-871b-58f5e69aa834, please check neutron logs for more information. [ 602.986455] env[63088]: ERROR nova.compute.manager [ 602.986455] env[63088]: Traceback (most recent call last): [ 602.986455] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 602.986455] env[63088]: listener.cb(fileno) [ 602.986455] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 602.986455] env[63088]: result = function(*args, **kwargs) [ 602.986455] env[63088]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 602.986455] env[63088]: return func(*args, **kwargs) [ 602.986455] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 602.986455] env[63088]: raise e [ 602.986455] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 602.986455] env[63088]: nwinfo = self.network_api.allocate_for_instance( [ 602.986455] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 602.986455] env[63088]: created_port_ids = self._update_ports_for_instance( [ 602.986455] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 602.986455] env[63088]: with excutils.save_and_reraise_exception(): [ 602.986455] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 602.986455] env[63088]: self.force_reraise() [ 602.986455] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 602.986455] env[63088]: raise self.value [ 602.986455] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 602.986455] env[63088]: updated_port = self._update_port( [ 602.986455] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 602.986455] env[63088]: _ensure_no_port_binding_failure(port) [ 602.986455] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 602.986455] env[63088]: raise exception.PortBindingFailed(port_id=port['id']) [ 602.987400] env[63088]: nova.exception.PortBindingFailed: Binding failed for port 203eaeff-191a-4386-871b-58f5e69aa834, please check neutron logs for more information. [ 602.987400] env[63088]: Removing descriptor: 17 [ 602.987400] env[63088]: ERROR nova.compute.manager [None req-79832ead-48a0-4aad-a47e-8e38c2a687da tempest-ServersTestJSON-106593364 tempest-ServersTestJSON-106593364-project-member] [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 203eaeff-191a-4386-871b-58f5e69aa834, please check neutron logs for more information. [ 602.987400] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] Traceback (most recent call last): [ 602.987400] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 602.987400] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] yield resources [ 602.987400] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 602.987400] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] self.driver.spawn(context, instance, image_meta, [ 602.987400] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 602.987400] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] self._vmops.spawn(context, instance, image_meta, injected_files, [ 602.987400] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 602.987400] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] vm_ref = self.build_virtual_machine(instance, [ 602.987775] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 602.987775] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] vif_infos = vmwarevif.get_vif_info(self._session, [ 602.987775] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 602.987775] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] for vif in network_info: [ 602.987775] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 602.987775] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] return self._sync_wrapper(fn, *args, **kwargs) [ 602.987775] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 602.987775] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] self.wait() [ 602.987775] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 602.987775] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] self[:] = self._gt.wait() [ 602.987775] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 602.987775] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] return self._exit_event.wait() [ 602.987775] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 602.988267] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] result = hub.switch() [ 602.988267] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 602.988267] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] return self.greenlet.switch() [ 602.988267] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 602.988267] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] result = function(*args, **kwargs) [ 602.988267] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 602.988267] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] return func(*args, **kwargs) [ 602.988267] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 602.988267] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] raise e [ 602.988267] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 602.988267] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] nwinfo = self.network_api.allocate_for_instance( [ 602.988267] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 602.988267] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] created_port_ids = self._update_ports_for_instance( [ 602.989324] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 602.989324] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] with excutils.save_and_reraise_exception(): [ 602.989324] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 602.989324] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] self.force_reraise() [ 602.989324] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 602.989324] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] raise self.value [ 602.989324] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 602.989324] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] updated_port = self._update_port( [ 602.989324] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 602.989324] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] _ensure_no_port_binding_failure(port) [ 602.989324] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 602.989324] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] raise exception.PortBindingFailed(port_id=port['id']) [ 602.989661] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] nova.exception.PortBindingFailed: Binding failed for port 203eaeff-191a-4386-871b-58f5e69aa834, please check neutron logs for more information. [ 602.989661] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] [ 602.989661] env[63088]: INFO nova.compute.manager [None req-79832ead-48a0-4aad-a47e-8e38c2a687da tempest-ServersTestJSON-106593364 tempest-ServersTestJSON-106593364-project-member] [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] Terminating instance [ 602.989661] env[63088]: DEBUG oslo_concurrency.lockutils [None req-79832ead-48a0-4aad-a47e-8e38c2a687da tempest-ServersTestJSON-106593364 tempest-ServersTestJSON-106593364-project-member] Acquiring lock "refresh_cache-34c01c2c-519b-44a1-aba2-c26e2c0b4737" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 602.989661] env[63088]: DEBUG oslo_concurrency.lockutils [None req-79832ead-48a0-4aad-a47e-8e38c2a687da tempest-ServersTestJSON-106593364 tempest-ServersTestJSON-106593364-project-member] Acquired lock "refresh_cache-34c01c2c-519b-44a1-aba2-c26e2c0b4737" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 602.989661] env[63088]: DEBUG nova.network.neutron [None req-79832ead-48a0-4aad-a47e-8e38c2a687da tempest-ServersTestJSON-106593364 tempest-ServersTestJSON-106593364-project-member] [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 603.080271] env[63088]: DEBUG nova.compute.manager [req-27b93d07-7a48-44ea-a8e9-61bd703e83d8 req-b18463e5-4cf4-4939-8dda-38865fcc0a31 service nova] [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] Received event network-changed-203eaeff-191a-4386-871b-58f5e69aa834 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 603.080468] env[63088]: DEBUG nova.compute.manager [req-27b93d07-7a48-44ea-a8e9-61bd703e83d8 req-b18463e5-4cf4-4939-8dda-38865fcc0a31 service nova] [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] Refreshing instance network info cache due to event network-changed-203eaeff-191a-4386-871b-58f5e69aa834. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 603.080656] env[63088]: DEBUG oslo_concurrency.lockutils [req-27b93d07-7a48-44ea-a8e9-61bd703e83d8 req-b18463e5-4cf4-4939-8dda-38865fcc0a31 service nova] Acquiring lock "refresh_cache-34c01c2c-519b-44a1-aba2-c26e2c0b4737" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 603.125337] env[63088]: DEBUG nova.policy [None req-a9773ab2-f8b9-4102-8e41-5ee881a8637e tempest-FloatingIPsAssociationNegativeTestJSON-189613961 tempest-FloatingIPsAssociationNegativeTestJSON-189613961-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b806f8731c0b4d4689093861a5b64be0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd3b19515b82f4cafb14c017c26bd86e9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 603.372582] env[63088]: DEBUG nova.compute.manager [None req-a9773ab2-f8b9-4102-8e41-5ee881a8637e tempest-FloatingIPsAssociationNegativeTestJSON-189613961 tempest-FloatingIPsAssociationNegativeTestJSON-189613961-project-member] [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 603.383084] env[63088]: DEBUG oslo_concurrency.lockutils [None req-e2a589c2-79d5-4f3a-a528-7aa96bc4ee1b tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.020s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 603.384041] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bd71c985-753f-4eb9-b297-7d458ea844aa tempest-ImagesOneServerNegativeTestJSON-1975365234 tempest-ImagesOneServerNegativeTestJSON-1975365234-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.638s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 603.385582] env[63088]: INFO nova.compute.claims [None req-bd71c985-753f-4eb9-b297-7d458ea844aa tempest-ImagesOneServerNegativeTestJSON-1975365234 tempest-ImagesOneServerNegativeTestJSON-1975365234-project-member] [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 603.512287] env[63088]: DEBUG nova.network.neutron [None req-79832ead-48a0-4aad-a47e-8e38c2a687da tempest-ServersTestJSON-106593364 tempest-ServersTestJSON-106593364-project-member] [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 603.618888] env[63088]: DEBUG nova.network.neutron [None req-79832ead-48a0-4aad-a47e-8e38c2a687da tempest-ServersTestJSON-106593364 tempest-ServersTestJSON-106593364-project-member] [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 603.986062] env[63088]: DEBUG nova.network.neutron [None req-a9773ab2-f8b9-4102-8e41-5ee881a8637e tempest-FloatingIPsAssociationNegativeTestJSON-189613961 tempest-FloatingIPsAssociationNegativeTestJSON-189613961-project-member] [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] Successfully created port: 8d873b9d-f3a1-455e-aeb3-c68d1fa4fea7 {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 604.122157] env[63088]: DEBUG oslo_concurrency.lockutils [None req-79832ead-48a0-4aad-a47e-8e38c2a687da tempest-ServersTestJSON-106593364 tempest-ServersTestJSON-106593364-project-member] Releasing lock "refresh_cache-34c01c2c-519b-44a1-aba2-c26e2c0b4737" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 604.122599] env[63088]: DEBUG nova.compute.manager [None req-79832ead-48a0-4aad-a47e-8e38c2a687da tempest-ServersTestJSON-106593364 tempest-ServersTestJSON-106593364-project-member] [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 604.122786] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-79832ead-48a0-4aad-a47e-8e38c2a687da tempest-ServersTestJSON-106593364 tempest-ServersTestJSON-106593364-project-member] [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 604.123107] env[63088]: DEBUG oslo_concurrency.lockutils [req-27b93d07-7a48-44ea-a8e9-61bd703e83d8 req-b18463e5-4cf4-4939-8dda-38865fcc0a31 service nova] Acquired lock "refresh_cache-34c01c2c-519b-44a1-aba2-c26e2c0b4737" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 604.123271] env[63088]: DEBUG nova.network.neutron [req-27b93d07-7a48-44ea-a8e9-61bd703e83d8 req-b18463e5-4cf4-4939-8dda-38865fcc0a31 service nova] [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] Refreshing network info cache for port 203eaeff-191a-4386-871b-58f5e69aa834 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 604.124371] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-951d2f4d-e704-410b-be19-e33ef15d05de {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.142390] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f75f9ed-5546-455a-85de-205d235cba18 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.167400] env[63088]: WARNING nova.virt.vmwareapi.vmops [None req-79832ead-48a0-4aad-a47e-8e38c2a687da tempest-ServersTestJSON-106593364 tempest-ServersTestJSON-106593364-project-member] [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 34c01c2c-519b-44a1-aba2-c26e2c0b4737 could not be found. [ 604.170014] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-79832ead-48a0-4aad-a47e-8e38c2a687da tempest-ServersTestJSON-106593364 tempest-ServersTestJSON-106593364-project-member] [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 604.170014] env[63088]: INFO nova.compute.manager [None req-79832ead-48a0-4aad-a47e-8e38c2a687da tempest-ServersTestJSON-106593364 tempest-ServersTestJSON-106593364-project-member] [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] Took 0.05 seconds to destroy the instance on the hypervisor. [ 604.170014] env[63088]: DEBUG oslo.service.loopingcall [None req-79832ead-48a0-4aad-a47e-8e38c2a687da tempest-ServersTestJSON-106593364 tempest-ServersTestJSON-106593364-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 604.170014] env[63088]: DEBUG nova.compute.manager [-] [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 604.170014] env[63088]: DEBUG nova.network.neutron [-] [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 604.185042] env[63088]: DEBUG nova.network.neutron [-] [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 604.383234] env[63088]: DEBUG nova.compute.manager [None req-a9773ab2-f8b9-4102-8e41-5ee881a8637e tempest-FloatingIPsAssociationNegativeTestJSON-189613961 tempest-FloatingIPsAssociationNegativeTestJSON-189613961-project-member] [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 604.421340] env[63088]: DEBUG nova.virt.hardware [None req-a9773ab2-f8b9-4102-8e41-5ee881a8637e tempest-FloatingIPsAssociationNegativeTestJSON-189613961 tempest-FloatingIPsAssociationNegativeTestJSON-189613961-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 604.421588] env[63088]: DEBUG nova.virt.hardware [None req-a9773ab2-f8b9-4102-8e41-5ee881a8637e tempest-FloatingIPsAssociationNegativeTestJSON-189613961 tempest-FloatingIPsAssociationNegativeTestJSON-189613961-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 604.421744] env[63088]: DEBUG nova.virt.hardware [None req-a9773ab2-f8b9-4102-8e41-5ee881a8637e tempest-FloatingIPsAssociationNegativeTestJSON-189613961 tempest-FloatingIPsAssociationNegativeTestJSON-189613961-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 604.421922] env[63088]: DEBUG nova.virt.hardware [None req-a9773ab2-f8b9-4102-8e41-5ee881a8637e tempest-FloatingIPsAssociationNegativeTestJSON-189613961 tempest-FloatingIPsAssociationNegativeTestJSON-189613961-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 604.422080] env[63088]: DEBUG nova.virt.hardware [None req-a9773ab2-f8b9-4102-8e41-5ee881a8637e tempest-FloatingIPsAssociationNegativeTestJSON-189613961 tempest-FloatingIPsAssociationNegativeTestJSON-189613961-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 604.422234] env[63088]: DEBUG nova.virt.hardware [None req-a9773ab2-f8b9-4102-8e41-5ee881a8637e tempest-FloatingIPsAssociationNegativeTestJSON-189613961 tempest-FloatingIPsAssociationNegativeTestJSON-189613961-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 604.422440] env[63088]: DEBUG nova.virt.hardware [None req-a9773ab2-f8b9-4102-8e41-5ee881a8637e tempest-FloatingIPsAssociationNegativeTestJSON-189613961 tempest-FloatingIPsAssociationNegativeTestJSON-189613961-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 604.422597] env[63088]: DEBUG nova.virt.hardware [None req-a9773ab2-f8b9-4102-8e41-5ee881a8637e tempest-FloatingIPsAssociationNegativeTestJSON-189613961 tempest-FloatingIPsAssociationNegativeTestJSON-189613961-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 604.422761] env[63088]: DEBUG nova.virt.hardware [None req-a9773ab2-f8b9-4102-8e41-5ee881a8637e tempest-FloatingIPsAssociationNegativeTestJSON-189613961 tempest-FloatingIPsAssociationNegativeTestJSON-189613961-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 604.422919] env[63088]: DEBUG nova.virt.hardware [None req-a9773ab2-f8b9-4102-8e41-5ee881a8637e tempest-FloatingIPsAssociationNegativeTestJSON-189613961 tempest-FloatingIPsAssociationNegativeTestJSON-189613961-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 604.423104] env[63088]: DEBUG nova.virt.hardware [None req-a9773ab2-f8b9-4102-8e41-5ee881a8637e tempest-FloatingIPsAssociationNegativeTestJSON-189613961 tempest-FloatingIPsAssociationNegativeTestJSON-189613961-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 604.423967] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c107f7ba-564c-47e5-8f6f-9b629cc56dcb {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.436659] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bba19138-a4ab-43a9-9e55-efd7322ea2ae {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.656082] env[63088]: DEBUG nova.network.neutron [req-27b93d07-7a48-44ea-a8e9-61bd703e83d8 req-b18463e5-4cf4-4939-8dda-38865fcc0a31 service nova] [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 604.692174] env[63088]: DEBUG nova.network.neutron [-] [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 604.698593] env[63088]: DEBUG oslo_concurrency.lockutils [None req-7af44f64-0067-49a1-ac0c-3ee5b8047867 tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] Acquiring lock "f2b4485c-ebcb-4063-a25a-e2efe6d1758f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 604.698807] env[63088]: DEBUG oslo_concurrency.lockutils [None req-7af44f64-0067-49a1-ac0c-3ee5b8047867 tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] Lock "f2b4485c-ebcb-4063-a25a-e2efe6d1758f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 604.754371] env[63088]: DEBUG nova.network.neutron [req-27b93d07-7a48-44ea-a8e9-61bd703e83d8 req-b18463e5-4cf4-4939-8dda-38865fcc0a31 service nova] [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 604.884914] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b330d4a7-a4b3-434e-9437-b29723cb224f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.893602] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d86c860d-70ce-415e-bf90-01352828ad97 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.937364] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edfd557b-f067-4d31-8f2c-21af9ab61703 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.951984] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fe156b6-3c2b-48df-92db-85c58395698e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.967900] env[63088]: DEBUG nova.compute.provider_tree [None req-bd71c985-753f-4eb9-b297-7d458ea844aa tempest-ImagesOneServerNegativeTestJSON-1975365234 tempest-ImagesOneServerNegativeTestJSON-1975365234-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 605.199720] env[63088]: INFO nova.compute.manager [-] [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] Took 1.03 seconds to deallocate network for instance. [ 605.202938] env[63088]: DEBUG nova.compute.claims [None req-79832ead-48a0-4aad-a47e-8e38c2a687da tempest-ServersTestJSON-106593364 tempest-ServersTestJSON-106593364-project-member] [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] Aborting claim: {{(pid=63088) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 605.203417] env[63088]: DEBUG oslo_concurrency.lockutils [None req-79832ead-48a0-4aad-a47e-8e38c2a687da tempest-ServersTestJSON-106593364 tempest-ServersTestJSON-106593364-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 605.260533] env[63088]: DEBUG oslo_concurrency.lockutils [req-27b93d07-7a48-44ea-a8e9-61bd703e83d8 req-b18463e5-4cf4-4939-8dda-38865fcc0a31 service nova] Releasing lock "refresh_cache-34c01c2c-519b-44a1-aba2-c26e2c0b4737" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 605.423453] env[63088]: DEBUG nova.compute.manager [req-831d8e5d-1e58-4de9-933d-75f3ed73f617 req-0beadbdc-0bee-4db1-9cd6-7ead4745c914 service nova] [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] Received event network-vif-deleted-203eaeff-191a-4386-871b-58f5e69aa834 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 605.473070] env[63088]: DEBUG nova.scheduler.client.report [None req-bd71c985-753f-4eb9-b297-7d458ea844aa tempest-ImagesOneServerNegativeTestJSON-1975365234 tempest-ImagesOneServerNegativeTestJSON-1975365234-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 605.801056] env[63088]: DEBUG oslo_concurrency.lockutils [None req-03263d3e-caae-4f24-95aa-ac7aedd64259 tempest-ImagesOneServerTestJSON-44065314 tempest-ImagesOneServerTestJSON-44065314-project-member] Acquiring lock "c63c1740-fcec-47fa-a7b9-986b83c4abd9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 605.801342] env[63088]: DEBUG oslo_concurrency.lockutils [None req-03263d3e-caae-4f24-95aa-ac7aedd64259 tempest-ImagesOneServerTestJSON-44065314 tempest-ImagesOneServerTestJSON-44065314-project-member] Lock "c63c1740-fcec-47fa-a7b9-986b83c4abd9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 605.981390] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bd71c985-753f-4eb9-b297-7d458ea844aa tempest-ImagesOneServerNegativeTestJSON-1975365234 tempest-ImagesOneServerNegativeTestJSON-1975365234-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.597s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 605.981989] env[63088]: DEBUG nova.compute.manager [None req-bd71c985-753f-4eb9-b297-7d458ea844aa tempest-ImagesOneServerNegativeTestJSON-1975365234 tempest-ImagesOneServerNegativeTestJSON-1975365234-project-member] [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 605.986712] env[63088]: DEBUG oslo_concurrency.lockutils [None req-83cac04a-fc75-40e5-8cdf-832028033c8d tempest-ServersWithSpecificFlavorTestJSON-501634894 tempest-ServersWithSpecificFlavorTestJSON-501634894-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.176s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 605.988288] env[63088]: INFO nova.compute.claims [None req-83cac04a-fc75-40e5-8cdf-832028033c8d tempest-ServersWithSpecificFlavorTestJSON-501634894 tempest-ServersWithSpecificFlavorTestJSON-501634894-project-member] [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 606.137531] env[63088]: DEBUG nova.compute.manager [req-4ef6822a-32f1-48ed-b822-630b00edf4f6 req-c1968455-943d-4d9e-a09f-8f89b48c738d service nova] [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] Received event network-changed-8d873b9d-f3a1-455e-aeb3-c68d1fa4fea7 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 606.137759] env[63088]: DEBUG nova.compute.manager [req-4ef6822a-32f1-48ed-b822-630b00edf4f6 req-c1968455-943d-4d9e-a09f-8f89b48c738d service nova] [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] Refreshing instance network info cache due to event network-changed-8d873b9d-f3a1-455e-aeb3-c68d1fa4fea7. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 606.137965] env[63088]: DEBUG oslo_concurrency.lockutils [req-4ef6822a-32f1-48ed-b822-630b00edf4f6 req-c1968455-943d-4d9e-a09f-8f89b48c738d service nova] Acquiring lock "refresh_cache-9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 606.138256] env[63088]: DEBUG oslo_concurrency.lockutils [req-4ef6822a-32f1-48ed-b822-630b00edf4f6 req-c1968455-943d-4d9e-a09f-8f89b48c738d service nova] Acquired lock "refresh_cache-9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 606.138500] env[63088]: DEBUG nova.network.neutron [req-4ef6822a-32f1-48ed-b822-630b00edf4f6 req-c1968455-943d-4d9e-a09f-8f89b48c738d service nova] [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] Refreshing network info cache for port 8d873b9d-f3a1-455e-aeb3-c68d1fa4fea7 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 606.333026] env[63088]: ERROR nova.compute.manager [None req-a9773ab2-f8b9-4102-8e41-5ee881a8637e tempest-FloatingIPsAssociationNegativeTestJSON-189613961 tempest-FloatingIPsAssociationNegativeTestJSON-189613961-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8d873b9d-f3a1-455e-aeb3-c68d1fa4fea7, please check neutron logs for more information. [ 606.333026] env[63088]: ERROR nova.compute.manager Traceback (most recent call last): [ 606.333026] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 606.333026] env[63088]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 606.333026] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 606.333026] env[63088]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 606.333026] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 606.333026] env[63088]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 606.333026] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 606.333026] env[63088]: ERROR nova.compute.manager self.force_reraise() [ 606.333026] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 606.333026] env[63088]: ERROR nova.compute.manager raise self.value [ 606.333026] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 606.333026] env[63088]: ERROR nova.compute.manager updated_port = self._update_port( [ 606.333026] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 606.333026] env[63088]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 606.333929] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 606.333929] env[63088]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 606.333929] env[63088]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8d873b9d-f3a1-455e-aeb3-c68d1fa4fea7, please check neutron logs for more information. [ 606.333929] env[63088]: ERROR nova.compute.manager [ 606.333929] env[63088]: Traceback (most recent call last): [ 606.333929] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 606.333929] env[63088]: listener.cb(fileno) [ 606.333929] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 606.333929] env[63088]: result = function(*args, **kwargs) [ 606.333929] env[63088]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 606.333929] env[63088]: return func(*args, **kwargs) [ 606.333929] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 606.333929] env[63088]: raise e [ 606.333929] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 606.333929] env[63088]: nwinfo = self.network_api.allocate_for_instance( [ 606.333929] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 606.333929] env[63088]: created_port_ids = self._update_ports_for_instance( [ 606.333929] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 606.333929] env[63088]: with excutils.save_and_reraise_exception(): [ 606.333929] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 606.333929] env[63088]: self.force_reraise() [ 606.333929] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 606.333929] env[63088]: raise self.value [ 606.333929] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 606.333929] env[63088]: updated_port = self._update_port( [ 606.333929] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 606.333929] env[63088]: _ensure_no_port_binding_failure(port) [ 606.333929] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 606.333929] env[63088]: raise exception.PortBindingFailed(port_id=port['id']) [ 606.335354] env[63088]: nova.exception.PortBindingFailed: Binding failed for port 8d873b9d-f3a1-455e-aeb3-c68d1fa4fea7, please check neutron logs for more information. [ 606.335354] env[63088]: Removing descriptor: 15 [ 606.335354] env[63088]: ERROR nova.compute.manager [None req-a9773ab2-f8b9-4102-8e41-5ee881a8637e tempest-FloatingIPsAssociationNegativeTestJSON-189613961 tempest-FloatingIPsAssociationNegativeTestJSON-189613961-project-member] [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8d873b9d-f3a1-455e-aeb3-c68d1fa4fea7, please check neutron logs for more information. [ 606.335354] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] Traceback (most recent call last): [ 606.335354] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 606.335354] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] yield resources [ 606.335354] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 606.335354] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] self.driver.spawn(context, instance, image_meta, [ 606.335354] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 606.335354] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 606.335354] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 606.335354] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] vm_ref = self.build_virtual_machine(instance, [ 606.336588] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 606.336588] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] vif_infos = vmwarevif.get_vif_info(self._session, [ 606.336588] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 606.336588] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] for vif in network_info: [ 606.336588] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 606.336588] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] return self._sync_wrapper(fn, *args, **kwargs) [ 606.336588] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 606.336588] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] self.wait() [ 606.336588] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 606.336588] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] self[:] = self._gt.wait() [ 606.336588] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 606.336588] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] return self._exit_event.wait() [ 606.336588] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 606.337750] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] result = hub.switch() [ 606.337750] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 606.337750] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] return self.greenlet.switch() [ 606.337750] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 606.337750] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] result = function(*args, **kwargs) [ 606.337750] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 606.337750] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] return func(*args, **kwargs) [ 606.337750] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 606.337750] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] raise e [ 606.337750] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 606.337750] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] nwinfo = self.network_api.allocate_for_instance( [ 606.337750] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 606.337750] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] created_port_ids = self._update_ports_for_instance( [ 606.339257] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 606.339257] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] with excutils.save_and_reraise_exception(): [ 606.339257] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 606.339257] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] self.force_reraise() [ 606.339257] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 606.339257] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] raise self.value [ 606.339257] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 606.339257] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] updated_port = self._update_port( [ 606.339257] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 606.339257] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] _ensure_no_port_binding_failure(port) [ 606.339257] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 606.339257] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] raise exception.PortBindingFailed(port_id=port['id']) [ 606.339674] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] nova.exception.PortBindingFailed: Binding failed for port 8d873b9d-f3a1-455e-aeb3-c68d1fa4fea7, please check neutron logs for more information. [ 606.339674] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] [ 606.339674] env[63088]: INFO nova.compute.manager [None req-a9773ab2-f8b9-4102-8e41-5ee881a8637e tempest-FloatingIPsAssociationNegativeTestJSON-189613961 tempest-FloatingIPsAssociationNegativeTestJSON-189613961-project-member] [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] Terminating instance [ 606.339674] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a9773ab2-f8b9-4102-8e41-5ee881a8637e tempest-FloatingIPsAssociationNegativeTestJSON-189613961 tempest-FloatingIPsAssociationNegativeTestJSON-189613961-project-member] Acquiring lock "refresh_cache-9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 606.495323] env[63088]: DEBUG nova.compute.utils [None req-bd71c985-753f-4eb9-b297-7d458ea844aa tempest-ImagesOneServerNegativeTestJSON-1975365234 tempest-ImagesOneServerNegativeTestJSON-1975365234-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 606.500543] env[63088]: DEBUG nova.compute.manager [None req-bd71c985-753f-4eb9-b297-7d458ea844aa tempest-ImagesOneServerNegativeTestJSON-1975365234 tempest-ImagesOneServerNegativeTestJSON-1975365234-project-member] [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 606.500543] env[63088]: DEBUG nova.network.neutron [None req-bd71c985-753f-4eb9-b297-7d458ea844aa tempest-ImagesOneServerNegativeTestJSON-1975365234 tempest-ImagesOneServerNegativeTestJSON-1975365234-project-member] [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 606.554030] env[63088]: DEBUG nova.policy [None req-bd71c985-753f-4eb9-b297-7d458ea844aa tempest-ImagesOneServerNegativeTestJSON-1975365234 tempest-ImagesOneServerNegativeTestJSON-1975365234-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7725ea8eb5db4b84b37850a9e68b008b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a9cb6229d0954523a161fc5a2c84695b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 606.668519] env[63088]: DEBUG nova.network.neutron [req-4ef6822a-32f1-48ed-b822-630b00edf4f6 req-c1968455-943d-4d9e-a09f-8f89b48c738d service nova] [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 606.854228] env[63088]: DEBUG nova.network.neutron [None req-bd71c985-753f-4eb9-b297-7d458ea844aa tempest-ImagesOneServerNegativeTestJSON-1975365234 tempest-ImagesOneServerNegativeTestJSON-1975365234-project-member] [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] Successfully created port: c84089e5-c386-4714-aa1d-ef6063e26260 {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 606.904190] env[63088]: DEBUG nova.network.neutron [req-4ef6822a-32f1-48ed-b822-630b00edf4f6 req-c1968455-943d-4d9e-a09f-8f89b48c738d service nova] [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 607.003053] env[63088]: DEBUG nova.compute.manager [None req-bd71c985-753f-4eb9-b297-7d458ea844aa tempest-ImagesOneServerNegativeTestJSON-1975365234 tempest-ImagesOneServerNegativeTestJSON-1975365234-project-member] [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 607.409018] env[63088]: DEBUG oslo_concurrency.lockutils [req-4ef6822a-32f1-48ed-b822-630b00edf4f6 req-c1968455-943d-4d9e-a09f-8f89b48c738d service nova] Releasing lock "refresh_cache-9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 607.409018] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a9773ab2-f8b9-4102-8e41-5ee881a8637e tempest-FloatingIPsAssociationNegativeTestJSON-189613961 tempest-FloatingIPsAssociationNegativeTestJSON-189613961-project-member] Acquired lock "refresh_cache-9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 607.409018] env[63088]: DEBUG nova.network.neutron [None req-a9773ab2-f8b9-4102-8e41-5ee881a8637e tempest-FloatingIPsAssociationNegativeTestJSON-189613961 tempest-FloatingIPsAssociationNegativeTestJSON-189613961-project-member] [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 607.481774] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f58c1d8-a5ed-419d-9582-5f09c694ce8e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.491437] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66e03da6-86e3-491e-a4ba-457171656adc {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.537740] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3c064e9-f2bb-41e8-a763-fb2cf8e68d31 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.546161] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2836fdae-251c-40b0-9006-88e84c22e0ff {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.563084] env[63088]: DEBUG nova.compute.provider_tree [None req-83cac04a-fc75-40e5-8cdf-832028033c8d tempest-ServersWithSpecificFlavorTestJSON-501634894 tempest-ServersWithSpecificFlavorTestJSON-501634894-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 607.941571] env[63088]: DEBUG nova.network.neutron [None req-a9773ab2-f8b9-4102-8e41-5ee881a8637e tempest-FloatingIPsAssociationNegativeTestJSON-189613961 tempest-FloatingIPsAssociationNegativeTestJSON-189613961-project-member] [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 608.035357] env[63088]: ERROR nova.compute.manager [None req-bd71c985-753f-4eb9-b297-7d458ea844aa tempest-ImagesOneServerNegativeTestJSON-1975365234 tempest-ImagesOneServerNegativeTestJSON-1975365234-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c84089e5-c386-4714-aa1d-ef6063e26260, please check neutron logs for more information. [ 608.035357] env[63088]: ERROR nova.compute.manager Traceback (most recent call last): [ 608.035357] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 608.035357] env[63088]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 608.035357] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 608.035357] env[63088]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 608.035357] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 608.035357] env[63088]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 608.035357] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 608.035357] env[63088]: ERROR nova.compute.manager self.force_reraise() [ 608.035357] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 608.035357] env[63088]: ERROR nova.compute.manager raise self.value [ 608.035357] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 608.035357] env[63088]: ERROR nova.compute.manager updated_port = self._update_port( [ 608.035357] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 608.035357] env[63088]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 608.035904] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 608.035904] env[63088]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 608.035904] env[63088]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c84089e5-c386-4714-aa1d-ef6063e26260, please check neutron logs for more information. [ 608.035904] env[63088]: ERROR nova.compute.manager [ 608.036795] env[63088]: Traceback (most recent call last): [ 608.036917] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 608.036917] env[63088]: listener.cb(fileno) [ 608.037031] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 608.037031] env[63088]: result = function(*args, **kwargs) [ 608.037115] env[63088]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 608.037115] env[63088]: return func(*args, **kwargs) [ 608.037193] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 608.037193] env[63088]: raise e [ 608.037271] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 608.037271] env[63088]: nwinfo = self.network_api.allocate_for_instance( [ 608.037347] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 608.037347] env[63088]: created_port_ids = self._update_ports_for_instance( [ 608.037587] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 608.037587] env[63088]: with excutils.save_and_reraise_exception(): [ 608.037703] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 608.037703] env[63088]: self.force_reraise() [ 608.037771] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 608.037771] env[63088]: raise self.value [ 608.037964] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 608.037964] env[63088]: updated_port = self._update_port( [ 608.038219] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 608.038219] env[63088]: _ensure_no_port_binding_failure(port) [ 608.038219] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 608.038219] env[63088]: raise exception.PortBindingFailed(port_id=port['id']) [ 608.038219] env[63088]: nova.exception.PortBindingFailed: Binding failed for port c84089e5-c386-4714-aa1d-ef6063e26260, please check neutron logs for more information. [ 608.038219] env[63088]: Removing descriptor: 15 [ 608.046115] env[63088]: DEBUG nova.compute.manager [None req-bd71c985-753f-4eb9-b297-7d458ea844aa tempest-ImagesOneServerNegativeTestJSON-1975365234 tempest-ImagesOneServerNegativeTestJSON-1975365234-project-member] [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 608.064050] env[63088]: DEBUG nova.scheduler.client.report [None req-83cac04a-fc75-40e5-8cdf-832028033c8d tempest-ServersWithSpecificFlavorTestJSON-501634894 tempest-ServersWithSpecificFlavorTestJSON-501634894-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 608.083018] env[63088]: DEBUG nova.virt.hardware [None req-bd71c985-753f-4eb9-b297-7d458ea844aa tempest-ImagesOneServerNegativeTestJSON-1975365234 tempest-ImagesOneServerNegativeTestJSON-1975365234-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 608.083018] env[63088]: DEBUG nova.virt.hardware [None req-bd71c985-753f-4eb9-b297-7d458ea844aa tempest-ImagesOneServerNegativeTestJSON-1975365234 tempest-ImagesOneServerNegativeTestJSON-1975365234-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 608.083018] env[63088]: DEBUG nova.virt.hardware [None req-bd71c985-753f-4eb9-b297-7d458ea844aa tempest-ImagesOneServerNegativeTestJSON-1975365234 tempest-ImagesOneServerNegativeTestJSON-1975365234-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 608.083261] env[63088]: DEBUG nova.virt.hardware [None req-bd71c985-753f-4eb9-b297-7d458ea844aa tempest-ImagesOneServerNegativeTestJSON-1975365234 tempest-ImagesOneServerNegativeTestJSON-1975365234-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 608.083261] env[63088]: DEBUG nova.virt.hardware [None req-bd71c985-753f-4eb9-b297-7d458ea844aa tempest-ImagesOneServerNegativeTestJSON-1975365234 tempest-ImagesOneServerNegativeTestJSON-1975365234-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 608.083261] env[63088]: DEBUG nova.virt.hardware [None req-bd71c985-753f-4eb9-b297-7d458ea844aa tempest-ImagesOneServerNegativeTestJSON-1975365234 tempest-ImagesOneServerNegativeTestJSON-1975365234-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 608.083261] env[63088]: DEBUG nova.virt.hardware [None req-bd71c985-753f-4eb9-b297-7d458ea844aa tempest-ImagesOneServerNegativeTestJSON-1975365234 tempest-ImagesOneServerNegativeTestJSON-1975365234-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 608.083261] env[63088]: DEBUG nova.virt.hardware [None req-bd71c985-753f-4eb9-b297-7d458ea844aa tempest-ImagesOneServerNegativeTestJSON-1975365234 tempest-ImagesOneServerNegativeTestJSON-1975365234-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 608.083422] env[63088]: DEBUG nova.virt.hardware [None req-bd71c985-753f-4eb9-b297-7d458ea844aa tempest-ImagesOneServerNegativeTestJSON-1975365234 tempest-ImagesOneServerNegativeTestJSON-1975365234-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 608.083842] env[63088]: DEBUG nova.virt.hardware [None req-bd71c985-753f-4eb9-b297-7d458ea844aa tempest-ImagesOneServerNegativeTestJSON-1975365234 tempest-ImagesOneServerNegativeTestJSON-1975365234-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 608.084213] env[63088]: DEBUG nova.virt.hardware [None req-bd71c985-753f-4eb9-b297-7d458ea844aa tempest-ImagesOneServerNegativeTestJSON-1975365234 tempest-ImagesOneServerNegativeTestJSON-1975365234-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 608.085406] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09537937-6d3e-4e25-97a9-6efa5d55343b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.094183] env[63088]: DEBUG nova.network.neutron [None req-a9773ab2-f8b9-4102-8e41-5ee881a8637e tempest-FloatingIPsAssociationNegativeTestJSON-189613961 tempest-FloatingIPsAssociationNegativeTestJSON-189613961-project-member] [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 608.105849] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2a6183a-8e5e-4ff7-8aff-4c75ee765b6c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.125336] env[63088]: ERROR nova.compute.manager [None req-bd71c985-753f-4eb9-b297-7d458ea844aa tempest-ImagesOneServerNegativeTestJSON-1975365234 tempest-ImagesOneServerNegativeTestJSON-1975365234-project-member] [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c84089e5-c386-4714-aa1d-ef6063e26260, please check neutron logs for more information. [ 608.125336] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] Traceback (most recent call last): [ 608.125336] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 608.125336] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] yield resources [ 608.125336] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 608.125336] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] self.driver.spawn(context, instance, image_meta, [ 608.125336] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 608.125336] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 608.125336] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 608.125336] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] vm_ref = self.build_virtual_machine(instance, [ 608.125336] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 608.125693] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] vif_infos = vmwarevif.get_vif_info(self._session, [ 608.125693] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 608.125693] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] for vif in network_info: [ 608.125693] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 608.125693] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] return self._sync_wrapper(fn, *args, **kwargs) [ 608.125693] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 608.125693] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] self.wait() [ 608.125693] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 608.125693] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] self[:] = self._gt.wait() [ 608.125693] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 608.125693] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] return self._exit_event.wait() [ 608.125693] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 608.125693] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] current.throw(*self._exc) [ 608.126372] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 608.126372] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] result = function(*args, **kwargs) [ 608.126372] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 608.126372] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] return func(*args, **kwargs) [ 608.126372] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 608.126372] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] raise e [ 608.126372] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 608.126372] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] nwinfo = self.network_api.allocate_for_instance( [ 608.126372] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 608.126372] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] created_port_ids = self._update_ports_for_instance( [ 608.126372] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 608.126372] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] with excutils.save_and_reraise_exception(): [ 608.126372] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 608.126901] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] self.force_reraise() [ 608.126901] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 608.126901] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] raise self.value [ 608.126901] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 608.126901] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] updated_port = self._update_port( [ 608.126901] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 608.126901] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] _ensure_no_port_binding_failure(port) [ 608.126901] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 608.126901] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] raise exception.PortBindingFailed(port_id=port['id']) [ 608.126901] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] nova.exception.PortBindingFailed: Binding failed for port c84089e5-c386-4714-aa1d-ef6063e26260, please check neutron logs for more information. [ 608.126901] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] [ 608.126901] env[63088]: INFO nova.compute.manager [None req-bd71c985-753f-4eb9-b297-7d458ea844aa tempest-ImagesOneServerNegativeTestJSON-1975365234 tempest-ImagesOneServerNegativeTestJSON-1975365234-project-member] [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] Terminating instance [ 608.127486] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bd71c985-753f-4eb9-b297-7d458ea844aa tempest-ImagesOneServerNegativeTestJSON-1975365234 tempest-ImagesOneServerNegativeTestJSON-1975365234-project-member] Acquiring lock "refresh_cache-ca92ddb4-cac8-450a-9b7e-d09ad082354a" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 608.127486] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bd71c985-753f-4eb9-b297-7d458ea844aa tempest-ImagesOneServerNegativeTestJSON-1975365234 tempest-ImagesOneServerNegativeTestJSON-1975365234-project-member] Acquired lock "refresh_cache-ca92ddb4-cac8-450a-9b7e-d09ad082354a" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 608.127587] env[63088]: DEBUG nova.network.neutron [None req-bd71c985-753f-4eb9-b297-7d458ea844aa tempest-ImagesOneServerNegativeTestJSON-1975365234 tempest-ImagesOneServerNegativeTestJSON-1975365234-project-member] [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 608.196685] env[63088]: DEBUG nova.compute.manager [req-afd33781-d419-4207-8ed6-98b505a56dcf req-283ff895-9539-4fe9-b453-b0b68fe81733 service nova] [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] Received event network-vif-deleted-8d873b9d-f3a1-455e-aeb3-c68d1fa4fea7 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 608.196685] env[63088]: DEBUG nova.compute.manager [req-afd33781-d419-4207-8ed6-98b505a56dcf req-283ff895-9539-4fe9-b453-b0b68fe81733 service nova] [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] Received event network-changed-c84089e5-c386-4714-aa1d-ef6063e26260 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 608.196685] env[63088]: DEBUG nova.compute.manager [req-afd33781-d419-4207-8ed6-98b505a56dcf req-283ff895-9539-4fe9-b453-b0b68fe81733 service nova] [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] Refreshing instance network info cache due to event network-changed-c84089e5-c386-4714-aa1d-ef6063e26260. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 608.196685] env[63088]: DEBUG oslo_concurrency.lockutils [req-afd33781-d419-4207-8ed6-98b505a56dcf req-283ff895-9539-4fe9-b453-b0b68fe81733 service nova] Acquiring lock "refresh_cache-ca92ddb4-cac8-450a-9b7e-d09ad082354a" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 608.575397] env[63088]: DEBUG oslo_concurrency.lockutils [None req-83cac04a-fc75-40e5-8cdf-832028033c8d tempest-ServersWithSpecificFlavorTestJSON-501634894 tempest-ServersWithSpecificFlavorTestJSON-501634894-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.589s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 608.579425] env[63088]: DEBUG nova.compute.manager [None req-83cac04a-fc75-40e5-8cdf-832028033c8d tempest-ServersWithSpecificFlavorTestJSON-501634894 tempest-ServersWithSpecificFlavorTestJSON-501634894-project-member] [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 608.580914] env[63088]: DEBUG oslo_concurrency.lockutils [None req-acfde423-bb82-457b-a17c-3c463abca710 tempest-ServersTestFqdnHostnames-1568147653 tempest-ServersTestFqdnHostnames-1568147653-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.701s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 608.582639] env[63088]: INFO nova.compute.claims [None req-acfde423-bb82-457b-a17c-3c463abca710 tempest-ServersTestFqdnHostnames-1568147653 tempest-ServersTestFqdnHostnames-1568147653-project-member] [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 608.605227] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a9773ab2-f8b9-4102-8e41-5ee881a8637e tempest-FloatingIPsAssociationNegativeTestJSON-189613961 tempest-FloatingIPsAssociationNegativeTestJSON-189613961-project-member] Releasing lock "refresh_cache-9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 608.605227] env[63088]: DEBUG nova.compute.manager [None req-a9773ab2-f8b9-4102-8e41-5ee881a8637e tempest-FloatingIPsAssociationNegativeTestJSON-189613961 tempest-FloatingIPsAssociationNegativeTestJSON-189613961-project-member] [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 608.605227] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-a9773ab2-f8b9-4102-8e41-5ee881a8637e tempest-FloatingIPsAssociationNegativeTestJSON-189613961 tempest-FloatingIPsAssociationNegativeTestJSON-189613961-project-member] [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 608.605227] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8bb90f99-adfe-4977-9d2c-2c160eb297f8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.619922] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dcb5233-0833-4ff8-b101-ad9bdc45e241 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.650339] env[63088]: WARNING nova.virt.vmwareapi.vmops [None req-a9773ab2-f8b9-4102-8e41-5ee881a8637e tempest-FloatingIPsAssociationNegativeTestJSON-189613961 tempest-FloatingIPsAssociationNegativeTestJSON-189613961-project-member] [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4 could not be found. [ 608.652080] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-a9773ab2-f8b9-4102-8e41-5ee881a8637e tempest-FloatingIPsAssociationNegativeTestJSON-189613961 tempest-FloatingIPsAssociationNegativeTestJSON-189613961-project-member] [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 608.652080] env[63088]: INFO nova.compute.manager [None req-a9773ab2-f8b9-4102-8e41-5ee881a8637e tempest-FloatingIPsAssociationNegativeTestJSON-189613961 tempest-FloatingIPsAssociationNegativeTestJSON-189613961-project-member] [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] Took 0.05 seconds to destroy the instance on the hypervisor. [ 608.652080] env[63088]: DEBUG oslo.service.loopingcall [None req-a9773ab2-f8b9-4102-8e41-5ee881a8637e tempest-FloatingIPsAssociationNegativeTestJSON-189613961 tempest-FloatingIPsAssociationNegativeTestJSON-189613961-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 608.652080] env[63088]: DEBUG nova.compute.manager [-] [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 608.652080] env[63088]: DEBUG nova.network.neutron [-] [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 608.654242] env[63088]: DEBUG nova.network.neutron [None req-bd71c985-753f-4eb9-b297-7d458ea844aa tempest-ImagesOneServerNegativeTestJSON-1975365234 tempest-ImagesOneServerNegativeTestJSON-1975365234-project-member] [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 608.689479] env[63088]: DEBUG nova.network.neutron [-] [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 608.838296] env[63088]: DEBUG nova.network.neutron [None req-bd71c985-753f-4eb9-b297-7d458ea844aa tempest-ImagesOneServerNegativeTestJSON-1975365234 tempest-ImagesOneServerNegativeTestJSON-1975365234-project-member] [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 609.087224] env[63088]: DEBUG nova.compute.utils [None req-83cac04a-fc75-40e5-8cdf-832028033c8d tempest-ServersWithSpecificFlavorTestJSON-501634894 tempest-ServersWithSpecificFlavorTestJSON-501634894-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 609.090797] env[63088]: DEBUG nova.compute.manager [None req-83cac04a-fc75-40e5-8cdf-832028033c8d tempest-ServersWithSpecificFlavorTestJSON-501634894 tempest-ServersWithSpecificFlavorTestJSON-501634894-project-member] [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 609.090971] env[63088]: DEBUG nova.network.neutron [None req-83cac04a-fc75-40e5-8cdf-832028033c8d tempest-ServersWithSpecificFlavorTestJSON-501634894 tempest-ServersWithSpecificFlavorTestJSON-501634894-project-member] [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 609.181133] env[63088]: DEBUG nova.policy [None req-83cac04a-fc75-40e5-8cdf-832028033c8d tempest-ServersWithSpecificFlavorTestJSON-501634894 tempest-ServersWithSpecificFlavorTestJSON-501634894-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd0149dd85df04ae6be74cb56e88687ba', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd45392ba12d04e66be04ba5946077c9d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 609.193680] env[63088]: DEBUG nova.network.neutron [-] [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 609.345207] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bd71c985-753f-4eb9-b297-7d458ea844aa tempest-ImagesOneServerNegativeTestJSON-1975365234 tempest-ImagesOneServerNegativeTestJSON-1975365234-project-member] Releasing lock "refresh_cache-ca92ddb4-cac8-450a-9b7e-d09ad082354a" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 609.345692] env[63088]: DEBUG nova.compute.manager [None req-bd71c985-753f-4eb9-b297-7d458ea844aa tempest-ImagesOneServerNegativeTestJSON-1975365234 tempest-ImagesOneServerNegativeTestJSON-1975365234-project-member] [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 609.345882] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-bd71c985-753f-4eb9-b297-7d458ea844aa tempest-ImagesOneServerNegativeTestJSON-1975365234 tempest-ImagesOneServerNegativeTestJSON-1975365234-project-member] [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 609.346259] env[63088]: DEBUG oslo_concurrency.lockutils [req-afd33781-d419-4207-8ed6-98b505a56dcf req-283ff895-9539-4fe9-b453-b0b68fe81733 service nova] Acquired lock "refresh_cache-ca92ddb4-cac8-450a-9b7e-d09ad082354a" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 609.346434] env[63088]: DEBUG nova.network.neutron [req-afd33781-d419-4207-8ed6-98b505a56dcf req-283ff895-9539-4fe9-b453-b0b68fe81733 service nova] [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] Refreshing network info cache for port c84089e5-c386-4714-aa1d-ef6063e26260 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 609.347872] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e5edf094-a817-4b94-bc87-00f10192166a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.359833] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d2367cd-3d5c-4f22-8045-ef3d6035cc63 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.389564] env[63088]: WARNING nova.virt.vmwareapi.vmops [None req-bd71c985-753f-4eb9-b297-7d458ea844aa tempest-ImagesOneServerNegativeTestJSON-1975365234 tempest-ImagesOneServerNegativeTestJSON-1975365234-project-member] [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ca92ddb4-cac8-450a-9b7e-d09ad082354a could not be found. [ 609.389828] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-bd71c985-753f-4eb9-b297-7d458ea844aa tempest-ImagesOneServerNegativeTestJSON-1975365234 tempest-ImagesOneServerNegativeTestJSON-1975365234-project-member] [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 609.390035] env[63088]: INFO nova.compute.manager [None req-bd71c985-753f-4eb9-b297-7d458ea844aa tempest-ImagesOneServerNegativeTestJSON-1975365234 tempest-ImagesOneServerNegativeTestJSON-1975365234-project-member] [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] Took 0.04 seconds to destroy the instance on the hypervisor. [ 609.390434] env[63088]: DEBUG oslo.service.loopingcall [None req-bd71c985-753f-4eb9-b297-7d458ea844aa tempest-ImagesOneServerNegativeTestJSON-1975365234 tempest-ImagesOneServerNegativeTestJSON-1975365234-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 609.390567] env[63088]: DEBUG nova.compute.manager [-] [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 609.391104] env[63088]: DEBUG nova.network.neutron [-] [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 609.412741] env[63088]: DEBUG nova.network.neutron [-] [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 609.592534] env[63088]: DEBUG nova.compute.manager [None req-83cac04a-fc75-40e5-8cdf-832028033c8d tempest-ServersWithSpecificFlavorTestJSON-501634894 tempest-ServersWithSpecificFlavorTestJSON-501634894-project-member] [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 609.624092] env[63088]: DEBUG nova.network.neutron [None req-83cac04a-fc75-40e5-8cdf-832028033c8d tempest-ServersWithSpecificFlavorTestJSON-501634894 tempest-ServersWithSpecificFlavorTestJSON-501634894-project-member] [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] Successfully created port: 228daa45-7c0d-4ae7-a3b3-1887b319e67d {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 609.695346] env[63088]: INFO nova.compute.manager [-] [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] Took 1.04 seconds to deallocate network for instance. [ 609.702416] env[63088]: DEBUG nova.compute.claims [None req-a9773ab2-f8b9-4102-8e41-5ee881a8637e tempest-FloatingIPsAssociationNegativeTestJSON-189613961 tempest-FloatingIPsAssociationNegativeTestJSON-189613961-project-member] [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] Aborting claim: {{(pid=63088) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 609.702633] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a9773ab2-f8b9-4102-8e41-5ee881a8637e tempest-FloatingIPsAssociationNegativeTestJSON-189613961 tempest-FloatingIPsAssociationNegativeTestJSON-189613961-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 609.888592] env[63088]: DEBUG nova.network.neutron [req-afd33781-d419-4207-8ed6-98b505a56dcf req-283ff895-9539-4fe9-b453-b0b68fe81733 service nova] [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 609.917456] env[63088]: DEBUG nova.network.neutron [-] [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 610.050142] env[63088]: DEBUG nova.network.neutron [req-afd33781-d419-4207-8ed6-98b505a56dcf req-283ff895-9539-4fe9-b453-b0b68fe81733 service nova] [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 610.108085] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-481d094e-60ce-47d6-838d-612a7aca5c4b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.118678] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30af2ce9-efc1-4af1-b242-365799caabf1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.155853] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa1c55e6-4b2c-49ab-b013-836acb8593b1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.165944] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80181fd1-2c66-4035-8a47-35032791c5fe {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.183626] env[63088]: DEBUG nova.compute.provider_tree [None req-acfde423-bb82-457b-a17c-3c463abca710 tempest-ServersTestFqdnHostnames-1568147653 tempest-ServersTestFqdnHostnames-1568147653-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 610.422072] env[63088]: INFO nova.compute.manager [-] [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] Took 1.03 seconds to deallocate network for instance. [ 610.424356] env[63088]: DEBUG nova.compute.claims [None req-bd71c985-753f-4eb9-b297-7d458ea844aa tempest-ImagesOneServerNegativeTestJSON-1975365234 tempest-ImagesOneServerNegativeTestJSON-1975365234-project-member] [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] Aborting claim: {{(pid=63088) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 610.424625] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bd71c985-753f-4eb9-b297-7d458ea844aa tempest-ImagesOneServerNegativeTestJSON-1975365234 tempest-ImagesOneServerNegativeTestJSON-1975365234-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 610.553364] env[63088]: DEBUG oslo_concurrency.lockutils [req-afd33781-d419-4207-8ed6-98b505a56dcf req-283ff895-9539-4fe9-b453-b0b68fe81733 service nova] Releasing lock "refresh_cache-ca92ddb4-cac8-450a-9b7e-d09ad082354a" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 610.553566] env[63088]: DEBUG nova.compute.manager [req-afd33781-d419-4207-8ed6-98b505a56dcf req-283ff895-9539-4fe9-b453-b0b68fe81733 service nova] [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] Received event network-vif-deleted-c84089e5-c386-4714-aa1d-ef6063e26260 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 610.606195] env[63088]: DEBUG nova.compute.manager [None req-83cac04a-fc75-40e5-8cdf-832028033c8d tempest-ServersWithSpecificFlavorTestJSON-501634894 tempest-ServersWithSpecificFlavorTestJSON-501634894-project-member] [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 610.641738] env[63088]: DEBUG nova.virt.hardware [None req-83cac04a-fc75-40e5-8cdf-832028033c8d tempest-ServersWithSpecificFlavorTestJSON-501634894 tempest-ServersWithSpecificFlavorTestJSON-501634894-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:14:26Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='1248403581',id=32,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_0-58307805',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 610.641972] env[63088]: DEBUG nova.virt.hardware [None req-83cac04a-fc75-40e5-8cdf-832028033c8d tempest-ServersWithSpecificFlavorTestJSON-501634894 tempest-ServersWithSpecificFlavorTestJSON-501634894-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 610.642134] env[63088]: DEBUG nova.virt.hardware [None req-83cac04a-fc75-40e5-8cdf-832028033c8d tempest-ServersWithSpecificFlavorTestJSON-501634894 tempest-ServersWithSpecificFlavorTestJSON-501634894-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 610.642312] env[63088]: DEBUG nova.virt.hardware [None req-83cac04a-fc75-40e5-8cdf-832028033c8d tempest-ServersWithSpecificFlavorTestJSON-501634894 tempest-ServersWithSpecificFlavorTestJSON-501634894-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 610.642447] env[63088]: DEBUG nova.virt.hardware [None req-83cac04a-fc75-40e5-8cdf-832028033c8d tempest-ServersWithSpecificFlavorTestJSON-501634894 tempest-ServersWithSpecificFlavorTestJSON-501634894-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 610.642584] env[63088]: DEBUG nova.virt.hardware [None req-83cac04a-fc75-40e5-8cdf-832028033c8d tempest-ServersWithSpecificFlavorTestJSON-501634894 tempest-ServersWithSpecificFlavorTestJSON-501634894-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 610.642781] env[63088]: DEBUG nova.virt.hardware [None req-83cac04a-fc75-40e5-8cdf-832028033c8d tempest-ServersWithSpecificFlavorTestJSON-501634894 tempest-ServersWithSpecificFlavorTestJSON-501634894-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 610.643207] env[63088]: DEBUG nova.virt.hardware [None req-83cac04a-fc75-40e5-8cdf-832028033c8d tempest-ServersWithSpecificFlavorTestJSON-501634894 tempest-ServersWithSpecificFlavorTestJSON-501634894-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 610.646079] env[63088]: DEBUG nova.virt.hardware [None req-83cac04a-fc75-40e5-8cdf-832028033c8d tempest-ServersWithSpecificFlavorTestJSON-501634894 tempest-ServersWithSpecificFlavorTestJSON-501634894-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 610.646079] env[63088]: DEBUG nova.virt.hardware [None req-83cac04a-fc75-40e5-8cdf-832028033c8d tempest-ServersWithSpecificFlavorTestJSON-501634894 tempest-ServersWithSpecificFlavorTestJSON-501634894-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 610.646079] env[63088]: DEBUG nova.virt.hardware [None req-83cac04a-fc75-40e5-8cdf-832028033c8d tempest-ServersWithSpecificFlavorTestJSON-501634894 tempest-ServersWithSpecificFlavorTestJSON-501634894-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 610.646891] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-694ee238-ad36-45f7-a689-e79823fcf0f6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.657961] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9367ae8-33a7-4029-8e51-ee7c6e42e445 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.663394] env[63088]: DEBUG nova.compute.manager [req-9adeeeb6-ef9b-4ae0-8af2-7671319583e5 req-9e820529-f382-40fd-b6a8-58dc5a5387d0 service nova] [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] Received event network-changed-228daa45-7c0d-4ae7-a3b3-1887b319e67d {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 610.663692] env[63088]: DEBUG nova.compute.manager [req-9adeeeb6-ef9b-4ae0-8af2-7671319583e5 req-9e820529-f382-40fd-b6a8-58dc5a5387d0 service nova] [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] Refreshing instance network info cache due to event network-changed-228daa45-7c0d-4ae7-a3b3-1887b319e67d. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 610.663897] env[63088]: DEBUG oslo_concurrency.lockutils [req-9adeeeb6-ef9b-4ae0-8af2-7671319583e5 req-9e820529-f382-40fd-b6a8-58dc5a5387d0 service nova] Acquiring lock "refresh_cache-d6c76851-7059-40a4-a39b-e625c99dfcd9" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 610.664081] env[63088]: DEBUG oslo_concurrency.lockutils [req-9adeeeb6-ef9b-4ae0-8af2-7671319583e5 req-9e820529-f382-40fd-b6a8-58dc5a5387d0 service nova] Acquired lock "refresh_cache-d6c76851-7059-40a4-a39b-e625c99dfcd9" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 610.664281] env[63088]: DEBUG nova.network.neutron [req-9adeeeb6-ef9b-4ae0-8af2-7671319583e5 req-9e820529-f382-40fd-b6a8-58dc5a5387d0 service nova] [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] Refreshing network info cache for port 228daa45-7c0d-4ae7-a3b3-1887b319e67d {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 610.687827] env[63088]: DEBUG nova.scheduler.client.report [None req-acfde423-bb82-457b-a17c-3c463abca710 tempest-ServersTestFqdnHostnames-1568147653 tempest-ServersTestFqdnHostnames-1568147653-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 610.778530] env[63088]: ERROR nova.compute.manager [None req-83cac04a-fc75-40e5-8cdf-832028033c8d tempest-ServersWithSpecificFlavorTestJSON-501634894 tempest-ServersWithSpecificFlavorTestJSON-501634894-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 228daa45-7c0d-4ae7-a3b3-1887b319e67d, please check neutron logs for more information. [ 610.778530] env[63088]: ERROR nova.compute.manager Traceback (most recent call last): [ 610.778530] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 610.778530] env[63088]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 610.778530] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 610.778530] env[63088]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 610.778530] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 610.778530] env[63088]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 610.778530] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 610.778530] env[63088]: ERROR nova.compute.manager self.force_reraise() [ 610.778530] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 610.778530] env[63088]: ERROR nova.compute.manager raise self.value [ 610.778530] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 610.778530] env[63088]: ERROR nova.compute.manager updated_port = self._update_port( [ 610.778530] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 610.778530] env[63088]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 610.779157] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 610.779157] env[63088]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 610.779157] env[63088]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 228daa45-7c0d-4ae7-a3b3-1887b319e67d, please check neutron logs for more information. [ 610.779157] env[63088]: ERROR nova.compute.manager [ 610.779157] env[63088]: Traceback (most recent call last): [ 610.779157] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 610.779157] env[63088]: listener.cb(fileno) [ 610.779157] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 610.779157] env[63088]: result = function(*args, **kwargs) [ 610.779157] env[63088]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 610.779157] env[63088]: return func(*args, **kwargs) [ 610.779157] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 610.779157] env[63088]: raise e [ 610.779157] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 610.779157] env[63088]: nwinfo = self.network_api.allocate_for_instance( [ 610.779157] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 610.779157] env[63088]: created_port_ids = self._update_ports_for_instance( [ 610.779157] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 610.779157] env[63088]: with excutils.save_and_reraise_exception(): [ 610.779157] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 610.779157] env[63088]: self.force_reraise() [ 610.779157] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 610.779157] env[63088]: raise self.value [ 610.779157] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 610.779157] env[63088]: updated_port = self._update_port( [ 610.779157] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 610.779157] env[63088]: _ensure_no_port_binding_failure(port) [ 610.779157] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 610.779157] env[63088]: raise exception.PortBindingFailed(port_id=port['id']) [ 610.781794] env[63088]: nova.exception.PortBindingFailed: Binding failed for port 228daa45-7c0d-4ae7-a3b3-1887b319e67d, please check neutron logs for more information. [ 610.781794] env[63088]: Removing descriptor: 17 [ 610.781794] env[63088]: ERROR nova.compute.manager [None req-83cac04a-fc75-40e5-8cdf-832028033c8d tempest-ServersWithSpecificFlavorTestJSON-501634894 tempest-ServersWithSpecificFlavorTestJSON-501634894-project-member] [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 228daa45-7c0d-4ae7-a3b3-1887b319e67d, please check neutron logs for more information. [ 610.781794] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] Traceback (most recent call last): [ 610.781794] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 610.781794] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] yield resources [ 610.781794] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 610.781794] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] self.driver.spawn(context, instance, image_meta, [ 610.781794] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 610.781794] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 610.781794] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 610.781794] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] vm_ref = self.build_virtual_machine(instance, [ 610.782187] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 610.782187] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] vif_infos = vmwarevif.get_vif_info(self._session, [ 610.782187] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 610.782187] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] for vif in network_info: [ 610.782187] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 610.782187] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] return self._sync_wrapper(fn, *args, **kwargs) [ 610.782187] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 610.782187] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] self.wait() [ 610.782187] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 610.782187] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] self[:] = self._gt.wait() [ 610.782187] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 610.782187] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] return self._exit_event.wait() [ 610.782187] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 610.782729] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] result = hub.switch() [ 610.782729] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 610.782729] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] return self.greenlet.switch() [ 610.782729] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 610.782729] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] result = function(*args, **kwargs) [ 610.782729] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 610.782729] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] return func(*args, **kwargs) [ 610.782729] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 610.782729] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] raise e [ 610.782729] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 610.782729] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] nwinfo = self.network_api.allocate_for_instance( [ 610.782729] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 610.782729] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] created_port_ids = self._update_ports_for_instance( [ 610.783564] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 610.783564] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] with excutils.save_and_reraise_exception(): [ 610.783564] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 610.783564] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] self.force_reraise() [ 610.783564] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 610.783564] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] raise self.value [ 610.783564] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 610.783564] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] updated_port = self._update_port( [ 610.783564] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 610.783564] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] _ensure_no_port_binding_failure(port) [ 610.783564] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 610.783564] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] raise exception.PortBindingFailed(port_id=port['id']) [ 610.784168] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] nova.exception.PortBindingFailed: Binding failed for port 228daa45-7c0d-4ae7-a3b3-1887b319e67d, please check neutron logs for more information. [ 610.784168] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] [ 610.784168] env[63088]: INFO nova.compute.manager [None req-83cac04a-fc75-40e5-8cdf-832028033c8d tempest-ServersWithSpecificFlavorTestJSON-501634894 tempest-ServersWithSpecificFlavorTestJSON-501634894-project-member] [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] Terminating instance [ 610.784168] env[63088]: DEBUG oslo_concurrency.lockutils [None req-83cac04a-fc75-40e5-8cdf-832028033c8d tempest-ServersWithSpecificFlavorTestJSON-501634894 tempest-ServersWithSpecificFlavorTestJSON-501634894-project-member] Acquiring lock "refresh_cache-d6c76851-7059-40a4-a39b-e625c99dfcd9" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 611.192027] env[63088]: DEBUG oslo_concurrency.lockutils [None req-acfde423-bb82-457b-a17c-3c463abca710 tempest-ServersTestFqdnHostnames-1568147653 tempest-ServersTestFqdnHostnames-1568147653-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.611s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 611.192544] env[63088]: DEBUG nova.compute.manager [None req-acfde423-bb82-457b-a17c-3c463abca710 tempest-ServersTestFqdnHostnames-1568147653 tempest-ServersTestFqdnHostnames-1568147653-project-member] [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 611.199021] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f5f50761-0430-407c-84cc-7b2c9cf8683b tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.497s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 611.199021] env[63088]: INFO nova.compute.claims [None req-f5f50761-0430-407c-84cc-7b2c9cf8683b tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 611.329262] env[63088]: DEBUG nova.network.neutron [req-9adeeeb6-ef9b-4ae0-8af2-7671319583e5 req-9e820529-f382-40fd-b6a8-58dc5a5387d0 service nova] [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 611.416186] env[63088]: DEBUG nova.network.neutron [req-9adeeeb6-ef9b-4ae0-8af2-7671319583e5 req-9e820529-f382-40fd-b6a8-58dc5a5387d0 service nova] [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 611.697964] env[63088]: DEBUG nova.compute.utils [None req-acfde423-bb82-457b-a17c-3c463abca710 tempest-ServersTestFqdnHostnames-1568147653 tempest-ServersTestFqdnHostnames-1568147653-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 611.699365] env[63088]: DEBUG nova.compute.manager [None req-acfde423-bb82-457b-a17c-3c463abca710 tempest-ServersTestFqdnHostnames-1568147653 tempest-ServersTestFqdnHostnames-1568147653-project-member] [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 611.699534] env[63088]: DEBUG nova.network.neutron [None req-acfde423-bb82-457b-a17c-3c463abca710 tempest-ServersTestFqdnHostnames-1568147653 tempest-ServersTestFqdnHostnames-1568147653-project-member] [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 611.742164] env[63088]: DEBUG nova.policy [None req-acfde423-bb82-457b-a17c-3c463abca710 tempest-ServersTestFqdnHostnames-1568147653 tempest-ServersTestFqdnHostnames-1568147653-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a77a04a855af4ee8955078d1ca55e24e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c95bf10c3da049c49e1b87303b7ce57d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 611.921038] env[63088]: DEBUG oslo_concurrency.lockutils [req-9adeeeb6-ef9b-4ae0-8af2-7671319583e5 req-9e820529-f382-40fd-b6a8-58dc5a5387d0 service nova] Releasing lock "refresh_cache-d6c76851-7059-40a4-a39b-e625c99dfcd9" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 611.922362] env[63088]: DEBUG oslo_concurrency.lockutils [None req-83cac04a-fc75-40e5-8cdf-832028033c8d tempest-ServersWithSpecificFlavorTestJSON-501634894 tempest-ServersWithSpecificFlavorTestJSON-501634894-project-member] Acquired lock "refresh_cache-d6c76851-7059-40a4-a39b-e625c99dfcd9" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 611.922362] env[63088]: DEBUG nova.network.neutron [None req-83cac04a-fc75-40e5-8cdf-832028033c8d tempest-ServersWithSpecificFlavorTestJSON-501634894 tempest-ServersWithSpecificFlavorTestJSON-501634894-project-member] [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 612.051305] env[63088]: DEBUG nova.network.neutron [None req-acfde423-bb82-457b-a17c-3c463abca710 tempest-ServersTestFqdnHostnames-1568147653 tempest-ServersTestFqdnHostnames-1568147653-project-member] [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] Successfully created port: c5aeb7be-57db-42c6-81aa-bbcdcef2ce85 {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 612.203559] env[63088]: DEBUG nova.compute.manager [None req-acfde423-bb82-457b-a17c-3c463abca710 tempest-ServersTestFqdnHostnames-1568147653 tempest-ServersTestFqdnHostnames-1568147653-project-member] [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 612.453279] env[63088]: DEBUG nova.network.neutron [None req-83cac04a-fc75-40e5-8cdf-832028033c8d tempest-ServersWithSpecificFlavorTestJSON-501634894 tempest-ServersWithSpecificFlavorTestJSON-501634894-project-member] [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 612.615195] env[63088]: DEBUG nova.network.neutron [None req-83cac04a-fc75-40e5-8cdf-832028033c8d tempest-ServersWithSpecificFlavorTestJSON-501634894 tempest-ServersWithSpecificFlavorTestJSON-501634894-project-member] [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 612.640677] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da56bc06-76d6-4bfd-8b73-a69a7608ad2b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.649924] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71e091ac-29b6-470b-9a66-21fcb5ce659e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.687560] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b57f9b23-e98e-4842-a61f-bb54dcdfe05d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.698105] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e2b0568-e6fc-44df-8cb2-09e2d2531288 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.715303] env[63088]: DEBUG nova.compute.provider_tree [None req-f5f50761-0430-407c-84cc-7b2c9cf8683b tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 612.751673] env[63088]: DEBUG nova.compute.manager [req-59a330e5-c3c8-4129-91e0-a0a1ba21a912 req-c057cd36-54b1-4034-a8f3-098974fe1d89 service nova] [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] Received event network-vif-deleted-228daa45-7c0d-4ae7-a3b3-1887b319e67d {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 613.095802] env[63088]: ERROR nova.compute.manager [None req-acfde423-bb82-457b-a17c-3c463abca710 tempest-ServersTestFqdnHostnames-1568147653 tempest-ServersTestFqdnHostnames-1568147653-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c5aeb7be-57db-42c6-81aa-bbcdcef2ce85, please check neutron logs for more information. [ 613.095802] env[63088]: ERROR nova.compute.manager Traceback (most recent call last): [ 613.095802] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 613.095802] env[63088]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 613.095802] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 613.095802] env[63088]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 613.095802] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 613.095802] env[63088]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 613.095802] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 613.095802] env[63088]: ERROR nova.compute.manager self.force_reraise() [ 613.095802] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 613.095802] env[63088]: ERROR nova.compute.manager raise self.value [ 613.095802] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 613.095802] env[63088]: ERROR nova.compute.manager updated_port = self._update_port( [ 613.095802] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 613.095802] env[63088]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 613.096653] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 613.096653] env[63088]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 613.096653] env[63088]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c5aeb7be-57db-42c6-81aa-bbcdcef2ce85, please check neutron logs for more information. [ 613.096653] env[63088]: ERROR nova.compute.manager [ 613.096653] env[63088]: Traceback (most recent call last): [ 613.096653] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 613.096653] env[63088]: listener.cb(fileno) [ 613.096653] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 613.096653] env[63088]: result = function(*args, **kwargs) [ 613.096653] env[63088]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 613.096653] env[63088]: return func(*args, **kwargs) [ 613.096653] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 613.096653] env[63088]: raise e [ 613.096653] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 613.096653] env[63088]: nwinfo = self.network_api.allocate_for_instance( [ 613.096653] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 613.096653] env[63088]: created_port_ids = self._update_ports_for_instance( [ 613.096653] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 613.096653] env[63088]: with excutils.save_and_reraise_exception(): [ 613.096653] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 613.096653] env[63088]: self.force_reraise() [ 613.096653] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 613.096653] env[63088]: raise self.value [ 613.096653] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 613.096653] env[63088]: updated_port = self._update_port( [ 613.096653] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 613.096653] env[63088]: _ensure_no_port_binding_failure(port) [ 613.096653] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 613.096653] env[63088]: raise exception.PortBindingFailed(port_id=port['id']) [ 613.097534] env[63088]: nova.exception.PortBindingFailed: Binding failed for port c5aeb7be-57db-42c6-81aa-bbcdcef2ce85, please check neutron logs for more information. [ 613.097534] env[63088]: Removing descriptor: 17 [ 613.120648] env[63088]: DEBUG oslo_concurrency.lockutils [None req-83cac04a-fc75-40e5-8cdf-832028033c8d tempest-ServersWithSpecificFlavorTestJSON-501634894 tempest-ServersWithSpecificFlavorTestJSON-501634894-project-member] Releasing lock "refresh_cache-d6c76851-7059-40a4-a39b-e625c99dfcd9" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 613.121096] env[63088]: DEBUG nova.compute.manager [None req-83cac04a-fc75-40e5-8cdf-832028033c8d tempest-ServersWithSpecificFlavorTestJSON-501634894 tempest-ServersWithSpecificFlavorTestJSON-501634894-project-member] [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 613.121288] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-83cac04a-fc75-40e5-8cdf-832028033c8d tempest-ServersWithSpecificFlavorTestJSON-501634894 tempest-ServersWithSpecificFlavorTestJSON-501634894-project-member] [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 613.121587] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-811cc297-0e36-4817-823b-c81a6bebf7d3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.133300] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a51beef-d0d1-4148-9623-7a81f1ad8dc2 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.160922] env[63088]: WARNING nova.virt.vmwareapi.vmops [None req-83cac04a-fc75-40e5-8cdf-832028033c8d tempest-ServersWithSpecificFlavorTestJSON-501634894 tempest-ServersWithSpecificFlavorTestJSON-501634894-project-member] [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d6c76851-7059-40a4-a39b-e625c99dfcd9 could not be found. [ 613.161174] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-83cac04a-fc75-40e5-8cdf-832028033c8d tempest-ServersWithSpecificFlavorTestJSON-501634894 tempest-ServersWithSpecificFlavorTestJSON-501634894-project-member] [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 613.161370] env[63088]: INFO nova.compute.manager [None req-83cac04a-fc75-40e5-8cdf-832028033c8d tempest-ServersWithSpecificFlavorTestJSON-501634894 tempest-ServersWithSpecificFlavorTestJSON-501634894-project-member] [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] Took 0.04 seconds to destroy the instance on the hypervisor. [ 613.161616] env[63088]: DEBUG oslo.service.loopingcall [None req-83cac04a-fc75-40e5-8cdf-832028033c8d tempest-ServersWithSpecificFlavorTestJSON-501634894 tempest-ServersWithSpecificFlavorTestJSON-501634894-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 613.161853] env[63088]: DEBUG nova.compute.manager [-] [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 613.161934] env[63088]: DEBUG nova.network.neutron [-] [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 613.182607] env[63088]: DEBUG nova.network.neutron [-] [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 613.222372] env[63088]: DEBUG nova.scheduler.client.report [None req-f5f50761-0430-407c-84cc-7b2c9cf8683b tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 613.226431] env[63088]: DEBUG nova.compute.manager [None req-acfde423-bb82-457b-a17c-3c463abca710 tempest-ServersTestFqdnHostnames-1568147653 tempest-ServersTestFqdnHostnames-1568147653-project-member] [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 613.261204] env[63088]: DEBUG nova.virt.hardware [None req-acfde423-bb82-457b-a17c-3c463abca710 tempest-ServersTestFqdnHostnames-1568147653 tempest-ServersTestFqdnHostnames-1568147653-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 613.261473] env[63088]: DEBUG nova.virt.hardware [None req-acfde423-bb82-457b-a17c-3c463abca710 tempest-ServersTestFqdnHostnames-1568147653 tempest-ServersTestFqdnHostnames-1568147653-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 613.261628] env[63088]: DEBUG nova.virt.hardware [None req-acfde423-bb82-457b-a17c-3c463abca710 tempest-ServersTestFqdnHostnames-1568147653 tempest-ServersTestFqdnHostnames-1568147653-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 613.261935] env[63088]: DEBUG nova.virt.hardware [None req-acfde423-bb82-457b-a17c-3c463abca710 tempest-ServersTestFqdnHostnames-1568147653 tempest-ServersTestFqdnHostnames-1568147653-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 613.262033] env[63088]: DEBUG nova.virt.hardware [None req-acfde423-bb82-457b-a17c-3c463abca710 tempest-ServersTestFqdnHostnames-1568147653 tempest-ServersTestFqdnHostnames-1568147653-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 613.262203] env[63088]: DEBUG nova.virt.hardware [None req-acfde423-bb82-457b-a17c-3c463abca710 tempest-ServersTestFqdnHostnames-1568147653 tempest-ServersTestFqdnHostnames-1568147653-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 613.262365] env[63088]: DEBUG nova.virt.hardware [None req-acfde423-bb82-457b-a17c-3c463abca710 tempest-ServersTestFqdnHostnames-1568147653 tempest-ServersTestFqdnHostnames-1568147653-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 613.262518] env[63088]: DEBUG nova.virt.hardware [None req-acfde423-bb82-457b-a17c-3c463abca710 tempest-ServersTestFqdnHostnames-1568147653 tempest-ServersTestFqdnHostnames-1568147653-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 613.262680] env[63088]: DEBUG nova.virt.hardware [None req-acfde423-bb82-457b-a17c-3c463abca710 tempest-ServersTestFqdnHostnames-1568147653 tempest-ServersTestFqdnHostnames-1568147653-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 613.262833] env[63088]: DEBUG nova.virt.hardware [None req-acfde423-bb82-457b-a17c-3c463abca710 tempest-ServersTestFqdnHostnames-1568147653 tempest-ServersTestFqdnHostnames-1568147653-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 613.263012] env[63088]: DEBUG nova.virt.hardware [None req-acfde423-bb82-457b-a17c-3c463abca710 tempest-ServersTestFqdnHostnames-1568147653 tempest-ServersTestFqdnHostnames-1568147653-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 613.263886] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6940839-351c-4e5a-81e6-c6bf40f45c4c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.274030] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b06d5f4-938a-425a-b393-c95ef3ed0fe3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.288823] env[63088]: ERROR nova.compute.manager [None req-acfde423-bb82-457b-a17c-3c463abca710 tempest-ServersTestFqdnHostnames-1568147653 tempest-ServersTestFqdnHostnames-1568147653-project-member] [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c5aeb7be-57db-42c6-81aa-bbcdcef2ce85, please check neutron logs for more information. [ 613.288823] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] Traceback (most recent call last): [ 613.288823] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 613.288823] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] yield resources [ 613.288823] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 613.288823] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] self.driver.spawn(context, instance, image_meta, [ 613.288823] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 613.288823] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] self._vmops.spawn(context, instance, image_meta, injected_files, [ 613.288823] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 613.288823] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] vm_ref = self.build_virtual_machine(instance, [ 613.288823] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 613.289260] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] vif_infos = vmwarevif.get_vif_info(self._session, [ 613.289260] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 613.289260] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] for vif in network_info: [ 613.289260] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 613.289260] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] return self._sync_wrapper(fn, *args, **kwargs) [ 613.289260] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 613.289260] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] self.wait() [ 613.289260] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 613.289260] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] self[:] = self._gt.wait() [ 613.289260] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 613.289260] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] return self._exit_event.wait() [ 613.289260] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 613.289260] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] current.throw(*self._exc) [ 613.289741] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 613.289741] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] result = function(*args, **kwargs) [ 613.289741] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 613.289741] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] return func(*args, **kwargs) [ 613.289741] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 613.289741] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] raise e [ 613.289741] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 613.289741] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] nwinfo = self.network_api.allocate_for_instance( [ 613.289741] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 613.289741] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] created_port_ids = self._update_ports_for_instance( [ 613.289741] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 613.289741] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] with excutils.save_and_reraise_exception(): [ 613.289741] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 613.290190] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] self.force_reraise() [ 613.290190] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 613.290190] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] raise self.value [ 613.290190] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 613.290190] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] updated_port = self._update_port( [ 613.290190] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 613.290190] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] _ensure_no_port_binding_failure(port) [ 613.290190] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 613.290190] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] raise exception.PortBindingFailed(port_id=port['id']) [ 613.290190] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] nova.exception.PortBindingFailed: Binding failed for port c5aeb7be-57db-42c6-81aa-bbcdcef2ce85, please check neutron logs for more information. [ 613.290190] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] [ 613.290190] env[63088]: INFO nova.compute.manager [None req-acfde423-bb82-457b-a17c-3c463abca710 tempest-ServersTestFqdnHostnames-1568147653 tempest-ServersTestFqdnHostnames-1568147653-project-member] [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] Terminating instance [ 613.291583] env[63088]: DEBUG oslo_concurrency.lockutils [None req-acfde423-bb82-457b-a17c-3c463abca710 tempest-ServersTestFqdnHostnames-1568147653 tempest-ServersTestFqdnHostnames-1568147653-project-member] Acquiring lock "refresh_cache-91a09343-c2d7-4f5d-ae0a-4a123af92fba" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 613.291755] env[63088]: DEBUG oslo_concurrency.lockutils [None req-acfde423-bb82-457b-a17c-3c463abca710 tempest-ServersTestFqdnHostnames-1568147653 tempest-ServersTestFqdnHostnames-1568147653-project-member] Acquired lock "refresh_cache-91a09343-c2d7-4f5d-ae0a-4a123af92fba" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 613.291921] env[63088]: DEBUG nova.network.neutron [None req-acfde423-bb82-457b-a17c-3c463abca710 tempest-ServersTestFqdnHostnames-1568147653 tempest-ServersTestFqdnHostnames-1568147653-project-member] [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 613.687205] env[63088]: DEBUG nova.network.neutron [-] [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 613.730182] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f5f50761-0430-407c-84cc-7b2c9cf8683b tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.534s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 613.730796] env[63088]: DEBUG nova.compute.manager [None req-f5f50761-0430-407c-84cc-7b2c9cf8683b tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 613.733594] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d1272b04-ee5c-4e9e-b0bf-6ca9d0629aae tempest-VolumesAssistedSnapshotsTest-1004656420 tempest-VolumesAssistedSnapshotsTest-1004656420-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 24.680s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 613.814585] env[63088]: DEBUG nova.network.neutron [None req-acfde423-bb82-457b-a17c-3c463abca710 tempest-ServersTestFqdnHostnames-1568147653 tempest-ServersTestFqdnHostnames-1568147653-project-member] [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 613.904964] env[63088]: DEBUG nova.network.neutron [None req-acfde423-bb82-457b-a17c-3c463abca710 tempest-ServersTestFqdnHostnames-1568147653 tempest-ServersTestFqdnHostnames-1568147653-project-member] [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 614.188420] env[63088]: INFO nova.compute.manager [-] [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] Took 1.03 seconds to deallocate network for instance. [ 614.190623] env[63088]: DEBUG nova.compute.claims [None req-83cac04a-fc75-40e5-8cdf-832028033c8d tempest-ServersWithSpecificFlavorTestJSON-501634894 tempest-ServersWithSpecificFlavorTestJSON-501634894-project-member] [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] Aborting claim: {{(pid=63088) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 614.190826] env[63088]: DEBUG oslo_concurrency.lockutils [None req-83cac04a-fc75-40e5-8cdf-832028033c8d tempest-ServersWithSpecificFlavorTestJSON-501634894 tempest-ServersWithSpecificFlavorTestJSON-501634894-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 614.242133] env[63088]: DEBUG nova.compute.utils [None req-f5f50761-0430-407c-84cc-7b2c9cf8683b tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 614.242555] env[63088]: DEBUG nova.compute.manager [None req-f5f50761-0430-407c-84cc-7b2c9cf8683b tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 614.242728] env[63088]: DEBUG nova.network.neutron [None req-f5f50761-0430-407c-84cc-7b2c9cf8683b tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 614.328258] env[63088]: DEBUG nova.policy [None req-f5f50761-0430-407c-84cc-7b2c9cf8683b tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '637ba82799ba4c09a3ee901704675860', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7cf8918ced834fdfaa1bf2350b666ce0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 614.407366] env[63088]: DEBUG oslo_concurrency.lockutils [None req-acfde423-bb82-457b-a17c-3c463abca710 tempest-ServersTestFqdnHostnames-1568147653 tempest-ServersTestFqdnHostnames-1568147653-project-member] Releasing lock "refresh_cache-91a09343-c2d7-4f5d-ae0a-4a123af92fba" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 614.407853] env[63088]: DEBUG nova.compute.manager [None req-acfde423-bb82-457b-a17c-3c463abca710 tempest-ServersTestFqdnHostnames-1568147653 tempest-ServersTestFqdnHostnames-1568147653-project-member] [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 614.408087] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-acfde423-bb82-457b-a17c-3c463abca710 tempest-ServersTestFqdnHostnames-1568147653 tempest-ServersTestFqdnHostnames-1568147653-project-member] [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 614.408455] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3a65fefe-1b4d-4b4d-bcb4-1f12dfa07fc7 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.421982] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3aca0a7-3657-4179-9c84-3c27c89bdb22 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.460293] env[63088]: WARNING nova.virt.vmwareapi.vmops [None req-acfde423-bb82-457b-a17c-3c463abca710 tempest-ServersTestFqdnHostnames-1568147653 tempest-ServersTestFqdnHostnames-1568147653-project-member] [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 91a09343-c2d7-4f5d-ae0a-4a123af92fba could not be found. [ 614.460293] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-acfde423-bb82-457b-a17c-3c463abca710 tempest-ServersTestFqdnHostnames-1568147653 tempest-ServersTestFqdnHostnames-1568147653-project-member] [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 614.460411] env[63088]: INFO nova.compute.manager [None req-acfde423-bb82-457b-a17c-3c463abca710 tempest-ServersTestFqdnHostnames-1568147653 tempest-ServersTestFqdnHostnames-1568147653-project-member] [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] Took 0.05 seconds to destroy the instance on the hypervisor. [ 614.462132] env[63088]: DEBUG oslo.service.loopingcall [None req-acfde423-bb82-457b-a17c-3c463abca710 tempest-ServersTestFqdnHostnames-1568147653 tempest-ServersTestFqdnHostnames-1568147653-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 614.463279] env[63088]: DEBUG nova.compute.manager [-] [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 614.463379] env[63088]: DEBUG nova.network.neutron [-] [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 614.494160] env[63088]: DEBUG nova.network.neutron [-] [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 614.713948] env[63088]: DEBUG nova.network.neutron [None req-f5f50761-0430-407c-84cc-7b2c9cf8683b tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] Successfully created port: f6eeb038-e18c-4985-82b2-4297f9655e25 {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 614.749962] env[63088]: DEBUG nova.compute.manager [None req-f5f50761-0430-407c-84cc-7b2c9cf8683b tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 614.789068] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0aac2188-0bd1-4153-b4cb-0153bb30c6c3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.795667] env[63088]: DEBUG nova.compute.manager [req-c7149d77-6a8a-48e2-a141-2bbb20fe271b req-86b8d26b-9bc3-486c-a791-4d8b0e122b83 service nova] [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] Received event network-changed-c5aeb7be-57db-42c6-81aa-bbcdcef2ce85 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 614.795892] env[63088]: DEBUG nova.compute.manager [req-c7149d77-6a8a-48e2-a141-2bbb20fe271b req-86b8d26b-9bc3-486c-a791-4d8b0e122b83 service nova] [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] Refreshing instance network info cache due to event network-changed-c5aeb7be-57db-42c6-81aa-bbcdcef2ce85. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 614.796166] env[63088]: DEBUG oslo_concurrency.lockutils [req-c7149d77-6a8a-48e2-a141-2bbb20fe271b req-86b8d26b-9bc3-486c-a791-4d8b0e122b83 service nova] Acquiring lock "refresh_cache-91a09343-c2d7-4f5d-ae0a-4a123af92fba" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 614.796320] env[63088]: DEBUG oslo_concurrency.lockutils [req-c7149d77-6a8a-48e2-a141-2bbb20fe271b req-86b8d26b-9bc3-486c-a791-4d8b0e122b83 service nova] Acquired lock "refresh_cache-91a09343-c2d7-4f5d-ae0a-4a123af92fba" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 614.796607] env[63088]: DEBUG nova.network.neutron [req-c7149d77-6a8a-48e2-a141-2bbb20fe271b req-86b8d26b-9bc3-486c-a791-4d8b0e122b83 service nova] [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] Refreshing network info cache for port c5aeb7be-57db-42c6-81aa-bbcdcef2ce85 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 614.803803] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8b929d6-daa7-49b5-b242-6a4a2fe1aa94 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.837779] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd11e209-36ae-4eaf-a935-496b6e4166a2 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.846589] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62598b02-009c-418a-917e-62cd31effeaf {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.862028] env[63088]: DEBUG nova.compute.provider_tree [None req-d1272b04-ee5c-4e9e-b0bf-6ca9d0629aae tempest-VolumesAssistedSnapshotsTest-1004656420 tempest-VolumesAssistedSnapshotsTest-1004656420-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 614.996351] env[63088]: DEBUG nova.network.neutron [-] [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 615.319891] env[63088]: DEBUG nova.network.neutron [req-c7149d77-6a8a-48e2-a141-2bbb20fe271b req-86b8d26b-9bc3-486c-a791-4d8b0e122b83 service nova] [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 615.368253] env[63088]: DEBUG nova.scheduler.client.report [None req-d1272b04-ee5c-4e9e-b0bf-6ca9d0629aae tempest-VolumesAssistedSnapshotsTest-1004656420 tempest-VolumesAssistedSnapshotsTest-1004656420-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 615.429021] env[63088]: DEBUG nova.network.neutron [req-c7149d77-6a8a-48e2-a141-2bbb20fe271b req-86b8d26b-9bc3-486c-a791-4d8b0e122b83 service nova] [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 615.503895] env[63088]: INFO nova.compute.manager [-] [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] Took 1.04 seconds to deallocate network for instance. [ 615.504881] env[63088]: DEBUG nova.compute.claims [None req-acfde423-bb82-457b-a17c-3c463abca710 tempest-ServersTestFqdnHostnames-1568147653 tempest-ServersTestFqdnHostnames-1568147653-project-member] [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] Aborting claim: {{(pid=63088) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 615.505086] env[63088]: DEBUG oslo_concurrency.lockutils [None req-acfde423-bb82-457b-a17c-3c463abca710 tempest-ServersTestFqdnHostnames-1568147653 tempest-ServersTestFqdnHostnames-1568147653-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 615.760425] env[63088]: DEBUG nova.compute.manager [None req-f5f50761-0430-407c-84cc-7b2c9cf8683b tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 615.790271] env[63088]: DEBUG nova.virt.hardware [None req-f5f50761-0430-407c-84cc-7b2c9cf8683b tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 615.790555] env[63088]: DEBUG nova.virt.hardware [None req-f5f50761-0430-407c-84cc-7b2c9cf8683b tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 615.790756] env[63088]: DEBUG nova.virt.hardware [None req-f5f50761-0430-407c-84cc-7b2c9cf8683b tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 615.790987] env[63088]: DEBUG nova.virt.hardware [None req-f5f50761-0430-407c-84cc-7b2c9cf8683b tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 615.792527] env[63088]: DEBUG nova.virt.hardware [None req-f5f50761-0430-407c-84cc-7b2c9cf8683b tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 615.792747] env[63088]: DEBUG nova.virt.hardware [None req-f5f50761-0430-407c-84cc-7b2c9cf8683b tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 615.793039] env[63088]: DEBUG nova.virt.hardware [None req-f5f50761-0430-407c-84cc-7b2c9cf8683b tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 615.794453] env[63088]: DEBUG nova.virt.hardware [None req-f5f50761-0430-407c-84cc-7b2c9cf8683b tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 615.794453] env[63088]: DEBUG nova.virt.hardware [None req-f5f50761-0430-407c-84cc-7b2c9cf8683b tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 615.794453] env[63088]: DEBUG nova.virt.hardware [None req-f5f50761-0430-407c-84cc-7b2c9cf8683b tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 615.794453] env[63088]: DEBUG nova.virt.hardware [None req-f5f50761-0430-407c-84cc-7b2c9cf8683b tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 615.794681] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dceb4ac-1072-40ca-a983-3266c9fafb62 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.804911] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5212c548-d30c-4d8c-abc4-4ce12a5c3284 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.877696] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d1272b04-ee5c-4e9e-b0bf-6ca9d0629aae tempest-VolumesAssistedSnapshotsTest-1004656420 tempest-VolumesAssistedSnapshotsTest-1004656420-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.143s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 615.877696] env[63088]: ERROR nova.compute.manager [None req-d1272b04-ee5c-4e9e-b0bf-6ca9d0629aae tempest-VolumesAssistedSnapshotsTest-1004656420 tempest-VolumesAssistedSnapshotsTest-1004656420-project-member] [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e0dd550d-f2d8-47df-9706-a8988ec41e69, please check neutron logs for more information. [ 615.877696] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] Traceback (most recent call last): [ 615.877696] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 615.877696] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] self.driver.spawn(context, instance, image_meta, [ 615.877696] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 615.877696] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 615.877696] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 615.877696] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] vm_ref = self.build_virtual_machine(instance, [ 615.878101] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 615.878101] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] vif_infos = vmwarevif.get_vif_info(self._session, [ 615.878101] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 615.878101] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] for vif in network_info: [ 615.878101] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 615.878101] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] return self._sync_wrapper(fn, *args, **kwargs) [ 615.878101] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 615.878101] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] self.wait() [ 615.878101] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 615.878101] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] self[:] = self._gt.wait() [ 615.878101] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 615.878101] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] return self._exit_event.wait() [ 615.878101] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 615.878484] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] result = hub.switch() [ 615.878484] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 615.878484] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] return self.greenlet.switch() [ 615.878484] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 615.878484] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] result = function(*args, **kwargs) [ 615.878484] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 615.878484] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] return func(*args, **kwargs) [ 615.878484] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 615.878484] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] raise e [ 615.878484] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 615.878484] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] nwinfo = self.network_api.allocate_for_instance( [ 615.878484] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 615.878484] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] created_port_ids = self._update_ports_for_instance( [ 615.878862] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 615.878862] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] with excutils.save_and_reraise_exception(): [ 615.878862] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 615.878862] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] self.force_reraise() [ 615.878862] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 615.878862] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] raise self.value [ 615.878862] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 615.878862] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] updated_port = self._update_port( [ 615.878862] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 615.878862] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] _ensure_no_port_binding_failure(port) [ 615.878862] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 615.878862] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] raise exception.PortBindingFailed(port_id=port['id']) [ 615.879887] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] nova.exception.PortBindingFailed: Binding failed for port e0dd550d-f2d8-47df-9706-a8988ec41e69, please check neutron logs for more information. [ 615.879887] env[63088]: ERROR nova.compute.manager [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] [ 615.879887] env[63088]: DEBUG nova.compute.utils [None req-d1272b04-ee5c-4e9e-b0bf-6ca9d0629aae tempest-VolumesAssistedSnapshotsTest-1004656420 tempest-VolumesAssistedSnapshotsTest-1004656420-project-member] [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] Binding failed for port e0dd550d-f2d8-47df-9706-a8988ec41e69, please check neutron logs for more information. {{(pid=63088) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 615.879887] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 24.455s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 615.880091] env[63088]: DEBUG nova.objects.instance [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63088) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 615.883796] env[63088]: DEBUG nova.compute.manager [None req-d1272b04-ee5c-4e9e-b0bf-6ca9d0629aae tempest-VolumesAssistedSnapshotsTest-1004656420 tempest-VolumesAssistedSnapshotsTest-1004656420-project-member] [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] Build of instance d2c07175-b635-4900-9dab-9767e5f6bce4 was re-scheduled: Binding failed for port e0dd550d-f2d8-47df-9706-a8988ec41e69, please check neutron logs for more information. {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 615.884368] env[63088]: DEBUG nova.compute.manager [None req-d1272b04-ee5c-4e9e-b0bf-6ca9d0629aae tempest-VolumesAssistedSnapshotsTest-1004656420 tempest-VolumesAssistedSnapshotsTest-1004656420-project-member] [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] Unplugging VIFs for instance {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 615.884682] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d1272b04-ee5c-4e9e-b0bf-6ca9d0629aae tempest-VolumesAssistedSnapshotsTest-1004656420 tempest-VolumesAssistedSnapshotsTest-1004656420-project-member] Acquiring lock "refresh_cache-d2c07175-b635-4900-9dab-9767e5f6bce4" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 615.884906] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d1272b04-ee5c-4e9e-b0bf-6ca9d0629aae tempest-VolumesAssistedSnapshotsTest-1004656420 tempest-VolumesAssistedSnapshotsTest-1004656420-project-member] Acquired lock "refresh_cache-d2c07175-b635-4900-9dab-9767e5f6bce4" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 615.885143] env[63088]: DEBUG nova.network.neutron [None req-d1272b04-ee5c-4e9e-b0bf-6ca9d0629aae tempest-VolumesAssistedSnapshotsTest-1004656420 tempest-VolumesAssistedSnapshotsTest-1004656420-project-member] [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 615.927695] env[63088]: ERROR nova.compute.manager [None req-f5f50761-0430-407c-84cc-7b2c9cf8683b tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f6eeb038-e18c-4985-82b2-4297f9655e25, please check neutron logs for more information. [ 615.927695] env[63088]: ERROR nova.compute.manager Traceback (most recent call last): [ 615.927695] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 615.927695] env[63088]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 615.927695] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 615.927695] env[63088]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 615.927695] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 615.927695] env[63088]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 615.927695] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 615.927695] env[63088]: ERROR nova.compute.manager self.force_reraise() [ 615.927695] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 615.927695] env[63088]: ERROR nova.compute.manager raise self.value [ 615.927695] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 615.927695] env[63088]: ERROR nova.compute.manager updated_port = self._update_port( [ 615.927695] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 615.927695] env[63088]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 615.928298] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 615.928298] env[63088]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 615.928298] env[63088]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f6eeb038-e18c-4985-82b2-4297f9655e25, please check neutron logs for more information. [ 615.928298] env[63088]: ERROR nova.compute.manager [ 615.928298] env[63088]: Traceback (most recent call last): [ 615.928298] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 615.928298] env[63088]: listener.cb(fileno) [ 615.928298] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 615.928298] env[63088]: result = function(*args, **kwargs) [ 615.928298] env[63088]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 615.928298] env[63088]: return func(*args, **kwargs) [ 615.928298] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 615.928298] env[63088]: raise e [ 615.928298] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 615.928298] env[63088]: nwinfo = self.network_api.allocate_for_instance( [ 615.928298] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 615.928298] env[63088]: created_port_ids = self._update_ports_for_instance( [ 615.928298] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 615.928298] env[63088]: with excutils.save_and_reraise_exception(): [ 615.928298] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 615.928298] env[63088]: self.force_reraise() [ 615.928298] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 615.928298] env[63088]: raise self.value [ 615.928298] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 615.928298] env[63088]: updated_port = self._update_port( [ 615.928298] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 615.928298] env[63088]: _ensure_no_port_binding_failure(port) [ 615.928298] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 615.928298] env[63088]: raise exception.PortBindingFailed(port_id=port['id']) [ 615.929155] env[63088]: nova.exception.PortBindingFailed: Binding failed for port f6eeb038-e18c-4985-82b2-4297f9655e25, please check neutron logs for more information. [ 615.929155] env[63088]: Removing descriptor: 17 [ 615.929155] env[63088]: ERROR nova.compute.manager [None req-f5f50761-0430-407c-84cc-7b2c9cf8683b tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f6eeb038-e18c-4985-82b2-4297f9655e25, please check neutron logs for more information. [ 615.929155] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] Traceback (most recent call last): [ 615.929155] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 615.929155] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] yield resources [ 615.929155] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 615.929155] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] self.driver.spawn(context, instance, image_meta, [ 615.929155] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 615.929155] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] self._vmops.spawn(context, instance, image_meta, injected_files, [ 615.929155] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 615.929155] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] vm_ref = self.build_virtual_machine(instance, [ 615.929561] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 615.929561] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] vif_infos = vmwarevif.get_vif_info(self._session, [ 615.929561] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 615.929561] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] for vif in network_info: [ 615.929561] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 615.929561] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] return self._sync_wrapper(fn, *args, **kwargs) [ 615.929561] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 615.929561] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] self.wait() [ 615.929561] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 615.929561] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] self[:] = self._gt.wait() [ 615.929561] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 615.929561] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] return self._exit_event.wait() [ 615.929561] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 615.929939] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] result = hub.switch() [ 615.929939] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 615.929939] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] return self.greenlet.switch() [ 615.929939] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 615.929939] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] result = function(*args, **kwargs) [ 615.929939] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 615.929939] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] return func(*args, **kwargs) [ 615.929939] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 615.929939] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] raise e [ 615.929939] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 615.929939] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] nwinfo = self.network_api.allocate_for_instance( [ 615.929939] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 615.929939] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] created_port_ids = self._update_ports_for_instance( [ 615.930339] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 615.930339] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] with excutils.save_and_reraise_exception(): [ 615.930339] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 615.930339] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] self.force_reraise() [ 615.930339] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 615.930339] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] raise self.value [ 615.930339] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 615.930339] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] updated_port = self._update_port( [ 615.930339] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 615.930339] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] _ensure_no_port_binding_failure(port) [ 615.930339] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 615.930339] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] raise exception.PortBindingFailed(port_id=port['id']) [ 615.930750] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] nova.exception.PortBindingFailed: Binding failed for port f6eeb038-e18c-4985-82b2-4297f9655e25, please check neutron logs for more information. [ 615.930750] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] [ 615.930750] env[63088]: INFO nova.compute.manager [None req-f5f50761-0430-407c-84cc-7b2c9cf8683b tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] Terminating instance [ 615.936240] env[63088]: DEBUG oslo_concurrency.lockutils [req-c7149d77-6a8a-48e2-a141-2bbb20fe271b req-86b8d26b-9bc3-486c-a791-4d8b0e122b83 service nova] Releasing lock "refresh_cache-91a09343-c2d7-4f5d-ae0a-4a123af92fba" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 615.936240] env[63088]: DEBUG nova.compute.manager [req-c7149d77-6a8a-48e2-a141-2bbb20fe271b req-86b8d26b-9bc3-486c-a791-4d8b0e122b83 service nova] [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] Received event network-vif-deleted-c5aeb7be-57db-42c6-81aa-bbcdcef2ce85 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 615.937044] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f5f50761-0430-407c-84cc-7b2c9cf8683b tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] Acquiring lock "refresh_cache-f29f6a5e-8621-4802-a679-d5c9a8fb2461" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 615.937044] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f5f50761-0430-407c-84cc-7b2c9cf8683b tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] Acquired lock "refresh_cache-f29f6a5e-8621-4802-a679-d5c9a8fb2461" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 615.937044] env[63088]: DEBUG nova.network.neutron [None req-f5f50761-0430-407c-84cc-7b2c9cf8683b tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 616.416875] env[63088]: DEBUG nova.network.neutron [None req-d1272b04-ee5c-4e9e-b0bf-6ca9d0629aae tempest-VolumesAssistedSnapshotsTest-1004656420 tempest-VolumesAssistedSnapshotsTest-1004656420-project-member] [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 616.454379] env[63088]: DEBUG nova.network.neutron [None req-f5f50761-0430-407c-84cc-7b2c9cf8683b tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 616.533198] env[63088]: DEBUG nova.network.neutron [None req-d1272b04-ee5c-4e9e-b0bf-6ca9d0629aae tempest-VolumesAssistedSnapshotsTest-1004656420 tempest-VolumesAssistedSnapshotsTest-1004656420-project-member] [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 616.550694] env[63088]: DEBUG nova.network.neutron [None req-f5f50761-0430-407c-84cc-7b2c9cf8683b tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 616.819311] env[63088]: DEBUG nova.compute.manager [req-ce68fb34-8649-4a6a-bf50-424c5962f46c req-585654a9-3a73-450c-bc6a-9538d6f809e0 service nova] [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] Received event network-changed-f6eeb038-e18c-4985-82b2-4297f9655e25 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 616.819655] env[63088]: DEBUG nova.compute.manager [req-ce68fb34-8649-4a6a-bf50-424c5962f46c req-585654a9-3a73-450c-bc6a-9538d6f809e0 service nova] [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] Refreshing instance network info cache due to event network-changed-f6eeb038-e18c-4985-82b2-4297f9655e25. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 616.819767] env[63088]: DEBUG oslo_concurrency.lockutils [req-ce68fb34-8649-4a6a-bf50-424c5962f46c req-585654a9-3a73-450c-bc6a-9538d6f809e0 service nova] Acquiring lock "refresh_cache-f29f6a5e-8621-4802-a679-d5c9a8fb2461" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 616.888683] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4f48f189-259b-496a-9b0b-41755c01cd3a tempest-ServersAdmin275Test-260510338 tempest-ServersAdmin275Test-260510338-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.009s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 616.889753] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0e2f1d1d-62ce-4a58-953f-25113ba15c0f tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.558s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 616.889980] env[63088]: DEBUG nova.objects.instance [None req-0e2f1d1d-62ce-4a58-953f-25113ba15c0f tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Lazy-loading 'resources' on Instance uuid e8cfbce4-cfb2-463e-9f38-78a058dab6e1 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 617.035458] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d1272b04-ee5c-4e9e-b0bf-6ca9d0629aae tempest-VolumesAssistedSnapshotsTest-1004656420 tempest-VolumesAssistedSnapshotsTest-1004656420-project-member] Releasing lock "refresh_cache-d2c07175-b635-4900-9dab-9767e5f6bce4" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 617.035880] env[63088]: DEBUG nova.compute.manager [None req-d1272b04-ee5c-4e9e-b0bf-6ca9d0629aae tempest-VolumesAssistedSnapshotsTest-1004656420 tempest-VolumesAssistedSnapshotsTest-1004656420-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 617.036128] env[63088]: DEBUG nova.compute.manager [None req-d1272b04-ee5c-4e9e-b0bf-6ca9d0629aae tempest-VolumesAssistedSnapshotsTest-1004656420 tempest-VolumesAssistedSnapshotsTest-1004656420-project-member] [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 617.036320] env[63088]: DEBUG nova.network.neutron [None req-d1272b04-ee5c-4e9e-b0bf-6ca9d0629aae tempest-VolumesAssistedSnapshotsTest-1004656420 tempest-VolumesAssistedSnapshotsTest-1004656420-project-member] [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 617.051950] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f5f50761-0430-407c-84cc-7b2c9cf8683b tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] Releasing lock "refresh_cache-f29f6a5e-8621-4802-a679-d5c9a8fb2461" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 617.052378] env[63088]: DEBUG nova.compute.manager [None req-f5f50761-0430-407c-84cc-7b2c9cf8683b tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 617.052571] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-f5f50761-0430-407c-84cc-7b2c9cf8683b tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 617.052870] env[63088]: DEBUG oslo_concurrency.lockutils [req-ce68fb34-8649-4a6a-bf50-424c5962f46c req-585654a9-3a73-450c-bc6a-9538d6f809e0 service nova] Acquired lock "refresh_cache-f29f6a5e-8621-4802-a679-d5c9a8fb2461" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 617.053050] env[63088]: DEBUG nova.network.neutron [req-ce68fb34-8649-4a6a-bf50-424c5962f46c req-585654a9-3a73-450c-bc6a-9538d6f809e0 service nova] [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] Refreshing network info cache for port f6eeb038-e18c-4985-82b2-4297f9655e25 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 617.054123] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-37113dc6-2e1c-4d68-8c6d-92bfea643a12 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.058028] env[63088]: DEBUG nova.network.neutron [None req-d1272b04-ee5c-4e9e-b0bf-6ca9d0629aae tempest-VolumesAssistedSnapshotsTest-1004656420 tempest-VolumesAssistedSnapshotsTest-1004656420-project-member] [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 617.070786] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85da6c0a-30f7-4faa-b74a-c0c84115d641 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.097325] env[63088]: WARNING nova.virt.vmwareapi.vmops [None req-f5f50761-0430-407c-84cc-7b2c9cf8683b tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f29f6a5e-8621-4802-a679-d5c9a8fb2461 could not be found. [ 617.097581] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-f5f50761-0430-407c-84cc-7b2c9cf8683b tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 617.097682] env[63088]: INFO nova.compute.manager [None req-f5f50761-0430-407c-84cc-7b2c9cf8683b tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] Took 0.05 seconds to destroy the instance on the hypervisor. [ 617.097925] env[63088]: DEBUG oslo.service.loopingcall [None req-f5f50761-0430-407c-84cc-7b2c9cf8683b tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 617.098198] env[63088]: DEBUG nova.compute.manager [-] [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 617.098293] env[63088]: DEBUG nova.network.neutron [-] [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 617.112249] env[63088]: DEBUG nova.network.neutron [-] [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 617.564101] env[63088]: DEBUG nova.network.neutron [None req-d1272b04-ee5c-4e9e-b0bf-6ca9d0629aae tempest-VolumesAssistedSnapshotsTest-1004656420 tempest-VolumesAssistedSnapshotsTest-1004656420-project-member] [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 617.580403] env[63088]: DEBUG nova.network.neutron [req-ce68fb34-8649-4a6a-bf50-424c5962f46c req-585654a9-3a73-450c-bc6a-9538d6f809e0 service nova] [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 617.614912] env[63088]: DEBUG nova.network.neutron [-] [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 617.699037] env[63088]: DEBUG nova.network.neutron [req-ce68fb34-8649-4a6a-bf50-424c5962f46c req-585654a9-3a73-450c-bc6a-9538d6f809e0 service nova] [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 617.808266] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d14479b-b431-4737-a5b2-1b2d90c0ca83 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.816560] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07672ce1-8cc3-4cd5-be68-30a4f8192892 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.849165] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8043dc8e-bc49-4217-aa1f-c54f14a6d818 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.858053] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76e29c3a-7b61-400c-a124-451cb427cf28 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.875164] env[63088]: DEBUG nova.compute.provider_tree [None req-0e2f1d1d-62ce-4a58-953f-25113ba15c0f tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 618.067600] env[63088]: INFO nova.compute.manager [None req-d1272b04-ee5c-4e9e-b0bf-6ca9d0629aae tempest-VolumesAssistedSnapshotsTest-1004656420 tempest-VolumesAssistedSnapshotsTest-1004656420-project-member] [instance: d2c07175-b635-4900-9dab-9767e5f6bce4] Took 1.03 seconds to deallocate network for instance. [ 618.117942] env[63088]: INFO nova.compute.manager [-] [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] Took 1.02 seconds to deallocate network for instance. [ 618.121222] env[63088]: DEBUG nova.compute.claims [None req-f5f50761-0430-407c-84cc-7b2c9cf8683b tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] Aborting claim: {{(pid=63088) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 618.121419] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f5f50761-0430-407c-84cc-7b2c9cf8683b tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 618.201600] env[63088]: DEBUG oslo_concurrency.lockutils [req-ce68fb34-8649-4a6a-bf50-424c5962f46c req-585654a9-3a73-450c-bc6a-9538d6f809e0 service nova] Releasing lock "refresh_cache-f29f6a5e-8621-4802-a679-d5c9a8fb2461" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 618.201861] env[63088]: DEBUG nova.compute.manager [req-ce68fb34-8649-4a6a-bf50-424c5962f46c req-585654a9-3a73-450c-bc6a-9538d6f809e0 service nova] [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] Received event network-vif-deleted-f6eeb038-e18c-4985-82b2-4297f9655e25 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 618.378063] env[63088]: DEBUG nova.scheduler.client.report [None req-0e2f1d1d-62ce-4a58-953f-25113ba15c0f tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 618.886969] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0e2f1d1d-62ce-4a58-953f-25113ba15c0f tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.997s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 618.891075] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b1b885bf-cef9-4808-8f16-707ab04499b3 tempest-ServersTestManualDisk-1859582357 tempest-ServersTestManualDisk-1859582357-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.060s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 618.891755] env[63088]: INFO nova.compute.claims [None req-b1b885bf-cef9-4808-8f16-707ab04499b3 tempest-ServersTestManualDisk-1859582357 tempest-ServersTestManualDisk-1859582357-project-member] [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 618.912432] env[63088]: INFO nova.scheduler.client.report [None req-0e2f1d1d-62ce-4a58-953f-25113ba15c0f tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Deleted allocations for instance e8cfbce4-cfb2-463e-9f38-78a058dab6e1 [ 619.100772] env[63088]: INFO nova.scheduler.client.report [None req-d1272b04-ee5c-4e9e-b0bf-6ca9d0629aae tempest-VolumesAssistedSnapshotsTest-1004656420 tempest-VolumesAssistedSnapshotsTest-1004656420-project-member] Deleted allocations for instance d2c07175-b635-4900-9dab-9767e5f6bce4 [ 619.427924] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0e2f1d1d-62ce-4a58-953f-25113ba15c0f tempest-ServersAdmin275Test-1816874308 tempest-ServersAdmin275Test-1816874308-project-member] Lock "e8cfbce4-cfb2-463e-9f38-78a058dab6e1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.894s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 619.609339] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d1272b04-ee5c-4e9e-b0bf-6ca9d0629aae tempest-VolumesAssistedSnapshotsTest-1004656420 tempest-VolumesAssistedSnapshotsTest-1004656420-project-member] Lock "d2c07175-b635-4900-9dab-9767e5f6bce4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 62.822s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 620.112863] env[63088]: DEBUG nova.compute.manager [None req-a5e283fa-bcc1-4186-a4e3-1ae9f8d221af tempest-ServerMetadataTestJSON-859565163 tempest-ServerMetadataTestJSON-859565163-project-member] [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 620.311785] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33cc3fb5-8b64-4c27-a581-5e13e1e57f74 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.320997] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a716f9cb-38f4-4068-b5ba-5d41c8295a0a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.357553] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64efab04-6e40-42df-a066-632959e772a6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.363432] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a6077124-74e2-4777-9beb-9998c6275315 tempest-ServersTestBootFromVolume-2105176458 tempest-ServersTestBootFromVolume-2105176458-project-member] Acquiring lock "407db84f-3322-42bc-baac-5528e856af8d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 620.363692] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a6077124-74e2-4777-9beb-9998c6275315 tempest-ServersTestBootFromVolume-2105176458 tempest-ServersTestBootFromVolume-2105176458-project-member] Lock "407db84f-3322-42bc-baac-5528e856af8d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 620.369818] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e6e1127-9b79-46ba-b140-ade0a3837c82 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.384764] env[63088]: DEBUG nova.compute.provider_tree [None req-b1b885bf-cef9-4808-8f16-707ab04499b3 tempest-ServersTestManualDisk-1859582357 tempest-ServersTestManualDisk-1859582357-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 620.666448] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a5e283fa-bcc1-4186-a4e3-1ae9f8d221af tempest-ServerMetadataTestJSON-859565163 tempest-ServerMetadataTestJSON-859565163-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 620.888920] env[63088]: DEBUG nova.scheduler.client.report [None req-b1b885bf-cef9-4808-8f16-707ab04499b3 tempest-ServersTestManualDisk-1859582357 tempest-ServersTestManualDisk-1859582357-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 621.393996] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b1b885bf-cef9-4808-8f16-707ab04499b3 tempest-ServersTestManualDisk-1859582357 tempest-ServersTestManualDisk-1859582357-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.504s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 621.394547] env[63088]: DEBUG nova.compute.manager [None req-b1b885bf-cef9-4808-8f16-707ab04499b3 tempest-ServersTestManualDisk-1859582357 tempest-ServersTestManualDisk-1859582357-project-member] [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 621.397293] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a1a58940-45a1-49d0-85a7-c4e62f4fb6f2 tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.847s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 621.397518] env[63088]: DEBUG nova.objects.instance [None req-a1a58940-45a1-49d0-85a7-c4e62f4fb6f2 tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Lazy-loading 'resources' on Instance uuid c4777721-3f65-455f-9973-c1ed0732de34 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 621.903377] env[63088]: DEBUG nova.compute.utils [None req-b1b885bf-cef9-4808-8f16-707ab04499b3 tempest-ServersTestManualDisk-1859582357 tempest-ServersTestManualDisk-1859582357-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 621.905101] env[63088]: DEBUG nova.compute.manager [None req-b1b885bf-cef9-4808-8f16-707ab04499b3 tempest-ServersTestManualDisk-1859582357 tempest-ServersTestManualDisk-1859582357-project-member] [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 621.905365] env[63088]: DEBUG nova.network.neutron [None req-b1b885bf-cef9-4808-8f16-707ab04499b3 tempest-ServersTestManualDisk-1859582357 tempest-ServersTestManualDisk-1859582357-project-member] [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 621.968526] env[63088]: DEBUG nova.policy [None req-b1b885bf-cef9-4808-8f16-707ab04499b3 tempest-ServersTestManualDisk-1859582357 tempest-ServersTestManualDisk-1859582357-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8d97fcc778934ccb803864d9c3e4f682', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '102687a0fbc545e1b2beeb677d98cca1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 622.359079] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df42f669-2df8-4681-aa42-be261a326489 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.369629] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fcb0263-bc1d-4823-bcc2-a90e4defb9ed {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.408394] env[63088]: DEBUG nova.compute.manager [None req-b1b885bf-cef9-4808-8f16-707ab04499b3 tempest-ServersTestManualDisk-1859582357 tempest-ServersTestManualDisk-1859582357-project-member] [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 622.416233] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb4e3157-0c5a-44af-ac6c-0edde56c5446 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.427283] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b179fd61-4253-45e8-8e80-1debc7e89075 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.447824] env[63088]: DEBUG nova.compute.provider_tree [None req-a1a58940-45a1-49d0-85a7-c4e62f4fb6f2 tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 622.558065] env[63088]: DEBUG nova.network.neutron [None req-b1b885bf-cef9-4808-8f16-707ab04499b3 tempest-ServersTestManualDisk-1859582357 tempest-ServersTestManualDisk-1859582357-project-member] [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] Successfully created port: fc9be7f9-f0c9-43c8-8828-f85621944087 {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 622.950654] env[63088]: DEBUG nova.scheduler.client.report [None req-a1a58940-45a1-49d0-85a7-c4e62f4fb6f2 tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 623.423697] env[63088]: DEBUG nova.compute.manager [None req-b1b885bf-cef9-4808-8f16-707ab04499b3 tempest-ServersTestManualDisk-1859582357 tempest-ServersTestManualDisk-1859582357-project-member] [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 623.461477] env[63088]: DEBUG nova.virt.hardware [None req-b1b885bf-cef9-4808-8f16-707ab04499b3 tempest-ServersTestManualDisk-1859582357 tempest-ServersTestManualDisk-1859582357-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 623.461758] env[63088]: DEBUG nova.virt.hardware [None req-b1b885bf-cef9-4808-8f16-707ab04499b3 tempest-ServersTestManualDisk-1859582357 tempest-ServersTestManualDisk-1859582357-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 623.461862] env[63088]: DEBUG nova.virt.hardware [None req-b1b885bf-cef9-4808-8f16-707ab04499b3 tempest-ServersTestManualDisk-1859582357 tempest-ServersTestManualDisk-1859582357-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 623.462059] env[63088]: DEBUG nova.virt.hardware [None req-b1b885bf-cef9-4808-8f16-707ab04499b3 tempest-ServersTestManualDisk-1859582357 tempest-ServersTestManualDisk-1859582357-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 623.462199] env[63088]: DEBUG nova.virt.hardware [None req-b1b885bf-cef9-4808-8f16-707ab04499b3 tempest-ServersTestManualDisk-1859582357 tempest-ServersTestManualDisk-1859582357-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 623.462346] env[63088]: DEBUG nova.virt.hardware [None req-b1b885bf-cef9-4808-8f16-707ab04499b3 tempest-ServersTestManualDisk-1859582357 tempest-ServersTestManualDisk-1859582357-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 623.462554] env[63088]: DEBUG nova.virt.hardware [None req-b1b885bf-cef9-4808-8f16-707ab04499b3 tempest-ServersTestManualDisk-1859582357 tempest-ServersTestManualDisk-1859582357-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 623.462982] env[63088]: DEBUG nova.virt.hardware [None req-b1b885bf-cef9-4808-8f16-707ab04499b3 tempest-ServersTestManualDisk-1859582357 tempest-ServersTestManualDisk-1859582357-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 623.462982] env[63088]: DEBUG nova.virt.hardware [None req-b1b885bf-cef9-4808-8f16-707ab04499b3 tempest-ServersTestManualDisk-1859582357 tempest-ServersTestManualDisk-1859582357-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 623.463094] env[63088]: DEBUG nova.virt.hardware [None req-b1b885bf-cef9-4808-8f16-707ab04499b3 tempest-ServersTestManualDisk-1859582357 tempest-ServersTestManualDisk-1859582357-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 623.463245] env[63088]: DEBUG nova.virt.hardware [None req-b1b885bf-cef9-4808-8f16-707ab04499b3 tempest-ServersTestManualDisk-1859582357 tempest-ServersTestManualDisk-1859582357-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 623.464070] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a1a58940-45a1-49d0-85a7-c4e62f4fb6f2 tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.067s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 623.466596] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dc7d943-70e4-4e11-9589-3b84ca1b3d5e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.469387] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0564aade-729d-4cf1-a59d-513384386c54 tempest-AttachInterfacesV270Test-1783457293 tempest-AttachInterfacesV270Test-1783457293-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.419s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 623.471458] env[63088]: INFO nova.compute.claims [None req-0564aade-729d-4cf1-a59d-513384386c54 tempest-AttachInterfacesV270Test-1783457293 tempest-AttachInterfacesV270Test-1783457293-project-member] [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 623.480741] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a951e32-5b3a-45c4-b0f6-590cb4724c1e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.502823] env[63088]: INFO nova.scheduler.client.report [None req-a1a58940-45a1-49d0-85a7-c4e62f4fb6f2 tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Deleted allocations for instance c4777721-3f65-455f-9973-c1ed0732de34 [ 624.013725] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a1a58940-45a1-49d0-85a7-c4e62f4fb6f2 tempest-ServerDiagnosticsV248Test-1179113054 tempest-ServerDiagnosticsV248Test-1179113054-project-member] Lock "c4777721-3f65-455f-9973-c1ed0732de34" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.543s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 624.055086] env[63088]: DEBUG nova.compute.manager [req-be7bf853-d0a6-46dd-b4be-8038a3724c63 req-b72aaae1-d962-49db-8a28-2ff8f22f4b89 service nova] [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] Received event network-changed-fc9be7f9-f0c9-43c8-8828-f85621944087 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 624.055086] env[63088]: DEBUG nova.compute.manager [req-be7bf853-d0a6-46dd-b4be-8038a3724c63 req-b72aaae1-d962-49db-8a28-2ff8f22f4b89 service nova] [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] Refreshing instance network info cache due to event network-changed-fc9be7f9-f0c9-43c8-8828-f85621944087. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 624.055086] env[63088]: DEBUG oslo_concurrency.lockutils [req-be7bf853-d0a6-46dd-b4be-8038a3724c63 req-b72aaae1-d962-49db-8a28-2ff8f22f4b89 service nova] Acquiring lock "refresh_cache-498c8e1e-18c6-4e09-a90d-106600bed2cd" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 624.055086] env[63088]: DEBUG oslo_concurrency.lockutils [req-be7bf853-d0a6-46dd-b4be-8038a3724c63 req-b72aaae1-d962-49db-8a28-2ff8f22f4b89 service nova] Acquired lock "refresh_cache-498c8e1e-18c6-4e09-a90d-106600bed2cd" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 624.056798] env[63088]: DEBUG nova.network.neutron [req-be7bf853-d0a6-46dd-b4be-8038a3724c63 req-b72aaae1-d962-49db-8a28-2ff8f22f4b89 service nova] [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] Refreshing network info cache for port fc9be7f9-f0c9-43c8-8828-f85621944087 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 624.230056] env[63088]: ERROR nova.compute.manager [None req-b1b885bf-cef9-4808-8f16-707ab04499b3 tempest-ServersTestManualDisk-1859582357 tempest-ServersTestManualDisk-1859582357-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port fc9be7f9-f0c9-43c8-8828-f85621944087, please check neutron logs for more information. [ 624.230056] env[63088]: ERROR nova.compute.manager Traceback (most recent call last): [ 624.230056] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 624.230056] env[63088]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 624.230056] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 624.230056] env[63088]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 624.230056] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 624.230056] env[63088]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 624.230056] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 624.230056] env[63088]: ERROR nova.compute.manager self.force_reraise() [ 624.230056] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 624.230056] env[63088]: ERROR nova.compute.manager raise self.value [ 624.230056] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 624.230056] env[63088]: ERROR nova.compute.manager updated_port = self._update_port( [ 624.230056] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 624.230056] env[63088]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 624.230486] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 624.230486] env[63088]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 624.230486] env[63088]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port fc9be7f9-f0c9-43c8-8828-f85621944087, please check neutron logs for more information. [ 624.230486] env[63088]: ERROR nova.compute.manager [ 624.230486] env[63088]: Traceback (most recent call last): [ 624.230486] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 624.230486] env[63088]: listener.cb(fileno) [ 624.230486] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 624.230486] env[63088]: result = function(*args, **kwargs) [ 624.230486] env[63088]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 624.230486] env[63088]: return func(*args, **kwargs) [ 624.230486] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 624.230486] env[63088]: raise e [ 624.230486] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 624.230486] env[63088]: nwinfo = self.network_api.allocate_for_instance( [ 624.230486] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 624.230486] env[63088]: created_port_ids = self._update_ports_for_instance( [ 624.230486] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 624.230486] env[63088]: with excutils.save_and_reraise_exception(): [ 624.230486] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 624.230486] env[63088]: self.force_reraise() [ 624.230486] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 624.230486] env[63088]: raise self.value [ 624.230486] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 624.230486] env[63088]: updated_port = self._update_port( [ 624.230486] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 624.230486] env[63088]: _ensure_no_port_binding_failure(port) [ 624.230486] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 624.230486] env[63088]: raise exception.PortBindingFailed(port_id=port['id']) [ 624.231176] env[63088]: nova.exception.PortBindingFailed: Binding failed for port fc9be7f9-f0c9-43c8-8828-f85621944087, please check neutron logs for more information. [ 624.231176] env[63088]: Removing descriptor: 15 [ 624.231176] env[63088]: ERROR nova.compute.manager [None req-b1b885bf-cef9-4808-8f16-707ab04499b3 tempest-ServersTestManualDisk-1859582357 tempest-ServersTestManualDisk-1859582357-project-member] [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port fc9be7f9-f0c9-43c8-8828-f85621944087, please check neutron logs for more information. [ 624.231176] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] Traceback (most recent call last): [ 624.231176] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 624.231176] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] yield resources [ 624.231176] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 624.231176] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] self.driver.spawn(context, instance, image_meta, [ 624.231176] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 624.231176] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] self._vmops.spawn(context, instance, image_meta, injected_files, [ 624.231176] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 624.231176] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] vm_ref = self.build_virtual_machine(instance, [ 624.231450] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 624.231450] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] vif_infos = vmwarevif.get_vif_info(self._session, [ 624.231450] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 624.231450] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] for vif in network_info: [ 624.231450] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 624.231450] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] return self._sync_wrapper(fn, *args, **kwargs) [ 624.231450] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 624.231450] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] self.wait() [ 624.231450] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 624.231450] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] self[:] = self._gt.wait() [ 624.231450] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 624.231450] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] return self._exit_event.wait() [ 624.231450] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 624.231721] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] result = hub.switch() [ 624.231721] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 624.231721] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] return self.greenlet.switch() [ 624.231721] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 624.231721] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] result = function(*args, **kwargs) [ 624.231721] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 624.231721] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] return func(*args, **kwargs) [ 624.231721] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 624.231721] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] raise e [ 624.231721] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 624.231721] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] nwinfo = self.network_api.allocate_for_instance( [ 624.231721] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 624.231721] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] created_port_ids = self._update_ports_for_instance( [ 624.232074] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 624.232074] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] with excutils.save_and_reraise_exception(): [ 624.232074] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 624.232074] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] self.force_reraise() [ 624.232074] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 624.232074] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] raise self.value [ 624.232074] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 624.232074] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] updated_port = self._update_port( [ 624.232074] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 624.232074] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] _ensure_no_port_binding_failure(port) [ 624.232074] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 624.232074] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] raise exception.PortBindingFailed(port_id=port['id']) [ 624.232388] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] nova.exception.PortBindingFailed: Binding failed for port fc9be7f9-f0c9-43c8-8828-f85621944087, please check neutron logs for more information. [ 624.232388] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] [ 624.232388] env[63088]: INFO nova.compute.manager [None req-b1b885bf-cef9-4808-8f16-707ab04499b3 tempest-ServersTestManualDisk-1859582357 tempest-ServersTestManualDisk-1859582357-project-member] [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] Terminating instance [ 624.235905] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b1b885bf-cef9-4808-8f16-707ab04499b3 tempest-ServersTestManualDisk-1859582357 tempest-ServersTestManualDisk-1859582357-project-member] Acquiring lock "refresh_cache-498c8e1e-18c6-4e09-a90d-106600bed2cd" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 624.756367] env[63088]: DEBUG nova.network.neutron [req-be7bf853-d0a6-46dd-b4be-8038a3724c63 req-b72aaae1-d962-49db-8a28-2ff8f22f4b89 service nova] [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 624.913176] env[63088]: DEBUG nova.network.neutron [req-be7bf853-d0a6-46dd-b4be-8038a3724c63 req-b72aaae1-d962-49db-8a28-2ff8f22f4b89 service nova] [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 624.981971] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c882743a-a412-4f42-b80e-8a6a7b74c432 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.991379] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88470e65-2498-499b-b5fa-581195f655ed {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.032782] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5609ae0-af28-4da2-b9cf-fe791ef6a5d2 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.042961] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b98f1a4-b1e5-4e98-a399-146a9ae9ee21 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.058611] env[63088]: DEBUG nova.compute.provider_tree [None req-0564aade-729d-4cf1-a59d-513384386c54 tempest-AttachInterfacesV270Test-1783457293 tempest-AttachInterfacesV270Test-1783457293-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 625.309355] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3b417ae5-c397-4c95-9150-090b3a6260b8 tempest-FloatingIPsAssociationTestJSON-911050413 tempest-FloatingIPsAssociationTestJSON-911050413-project-member] Acquiring lock "1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 625.309682] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3b417ae5-c397-4c95-9150-090b3a6260b8 tempest-FloatingIPsAssociationTestJSON-911050413 tempest-FloatingIPsAssociationTestJSON-911050413-project-member] Lock "1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 625.421048] env[63088]: DEBUG oslo_concurrency.lockutils [req-be7bf853-d0a6-46dd-b4be-8038a3724c63 req-b72aaae1-d962-49db-8a28-2ff8f22f4b89 service nova] Releasing lock "refresh_cache-498c8e1e-18c6-4e09-a90d-106600bed2cd" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 625.421048] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b1b885bf-cef9-4808-8f16-707ab04499b3 tempest-ServersTestManualDisk-1859582357 tempest-ServersTestManualDisk-1859582357-project-member] Acquired lock "refresh_cache-498c8e1e-18c6-4e09-a90d-106600bed2cd" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 625.421048] env[63088]: DEBUG nova.network.neutron [None req-b1b885bf-cef9-4808-8f16-707ab04499b3 tempest-ServersTestManualDisk-1859582357 tempest-ServersTestManualDisk-1859582357-project-member] [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 625.565739] env[63088]: DEBUG nova.scheduler.client.report [None req-0564aade-729d-4cf1-a59d-513384386c54 tempest-AttachInterfacesV270Test-1783457293 tempest-AttachInterfacesV270Test-1783457293-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 625.943849] env[63088]: DEBUG nova.network.neutron [None req-b1b885bf-cef9-4808-8f16-707ab04499b3 tempest-ServersTestManualDisk-1859582357 tempest-ServersTestManualDisk-1859582357-project-member] [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 626.073130] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0564aade-729d-4cf1-a59d-513384386c54 tempest-AttachInterfacesV270Test-1783457293 tempest-AttachInterfacesV270Test-1783457293-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.603s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 626.077590] env[63088]: DEBUG nova.compute.manager [None req-0564aade-729d-4cf1-a59d-513384386c54 tempest-AttachInterfacesV270Test-1783457293 tempest-AttachInterfacesV270Test-1783457293-project-member] [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 626.080350] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f81b332d-72eb-4149-a6a2-3c8a9c9bec53 tempest-ServersNegativeTestJSON-1008333314 tempest-ServersNegativeTestJSON-1008333314-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.107s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 626.082150] env[63088]: INFO nova.compute.claims [None req-f81b332d-72eb-4149-a6a2-3c8a9c9bec53 tempest-ServersNegativeTestJSON-1008333314 tempest-ServersNegativeTestJSON-1008333314-project-member] [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 626.097997] env[63088]: DEBUG nova.compute.manager [req-a222e8f1-baf1-4934-a3dd-f1d484e7238e req-e1da7b02-30e5-44a4-b1e0-0cf61a7a445d service nova] [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] Received event network-vif-deleted-fc9be7f9-f0c9-43c8-8828-f85621944087 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 626.111858] env[63088]: DEBUG nova.network.neutron [None req-b1b885bf-cef9-4808-8f16-707ab04499b3 tempest-ServersTestManualDisk-1859582357 tempest-ServersTestManualDisk-1859582357-project-member] [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 626.583480] env[63088]: DEBUG nova.compute.utils [None req-0564aade-729d-4cf1-a59d-513384386c54 tempest-AttachInterfacesV270Test-1783457293 tempest-AttachInterfacesV270Test-1783457293-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 626.584846] env[63088]: DEBUG nova.compute.manager [None req-0564aade-729d-4cf1-a59d-513384386c54 tempest-AttachInterfacesV270Test-1783457293 tempest-AttachInterfacesV270Test-1783457293-project-member] [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 626.587104] env[63088]: DEBUG nova.network.neutron [None req-0564aade-729d-4cf1-a59d-513384386c54 tempest-AttachInterfacesV270Test-1783457293 tempest-AttachInterfacesV270Test-1783457293-project-member] [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 626.618326] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b1b885bf-cef9-4808-8f16-707ab04499b3 tempest-ServersTestManualDisk-1859582357 tempest-ServersTestManualDisk-1859582357-project-member] Releasing lock "refresh_cache-498c8e1e-18c6-4e09-a90d-106600bed2cd" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 626.618326] env[63088]: DEBUG nova.compute.manager [None req-b1b885bf-cef9-4808-8f16-707ab04499b3 tempest-ServersTestManualDisk-1859582357 tempest-ServersTestManualDisk-1859582357-project-member] [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 626.618326] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-b1b885bf-cef9-4808-8f16-707ab04499b3 tempest-ServersTestManualDisk-1859582357 tempest-ServersTestManualDisk-1859582357-project-member] [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 626.618434] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-722dc5d4-de88-4151-ad16-283166f848b6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.629759] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4942cbb1-8707-4bdc-ade4-3b05840db687 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.658229] env[63088]: WARNING nova.virt.vmwareapi.vmops [None req-b1b885bf-cef9-4808-8f16-707ab04499b3 tempest-ServersTestManualDisk-1859582357 tempest-ServersTestManualDisk-1859582357-project-member] [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 498c8e1e-18c6-4e09-a90d-106600bed2cd could not be found. [ 626.658229] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-b1b885bf-cef9-4808-8f16-707ab04499b3 tempest-ServersTestManualDisk-1859582357 tempest-ServersTestManualDisk-1859582357-project-member] [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 626.658229] env[63088]: INFO nova.compute.manager [None req-b1b885bf-cef9-4808-8f16-707ab04499b3 tempest-ServersTestManualDisk-1859582357 tempest-ServersTestManualDisk-1859582357-project-member] [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] Took 0.04 seconds to destroy the instance on the hypervisor. [ 626.658229] env[63088]: DEBUG oslo.service.loopingcall [None req-b1b885bf-cef9-4808-8f16-707ab04499b3 tempest-ServersTestManualDisk-1859582357 tempest-ServersTestManualDisk-1859582357-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 626.658229] env[63088]: DEBUG nova.compute.manager [-] [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 626.658229] env[63088]: DEBUG nova.network.neutron [-] [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 626.664358] env[63088]: DEBUG nova.policy [None req-0564aade-729d-4cf1-a59d-513384386c54 tempest-AttachInterfacesV270Test-1783457293 tempest-AttachInterfacesV270Test-1783457293-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1c7ee138b22947d29b90d1c571abe2b6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '521511aaa80a422f9f0095629930a55a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 626.682066] env[63088]: DEBUG nova.network.neutron [-] [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 627.089448] env[63088]: DEBUG nova.compute.manager [None req-0564aade-729d-4cf1-a59d-513384386c54 tempest-AttachInterfacesV270Test-1783457293 tempest-AttachInterfacesV270Test-1783457293-project-member] [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 627.187090] env[63088]: DEBUG nova.network.neutron [None req-0564aade-729d-4cf1-a59d-513384386c54 tempest-AttachInterfacesV270Test-1783457293 tempest-AttachInterfacesV270Test-1783457293-project-member] [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] Successfully created port: c7b197ee-fb06-46ec-9aac-b7fb7874113c {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 627.189091] env[63088]: DEBUG nova.network.neutron [-] [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 627.570796] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a5a9e2c-0371-4c6f-81c1-7e77d86f7753 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.579937] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7bded99-9691-46c3-afa1-03e90a43fc99 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.615721] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-621cffd7-12b4-4690-87a1-6d190f000799 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.624820] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71d9851a-c0e3-497a-b93c-ca9b3b62b0f0 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.640327] env[63088]: DEBUG nova.compute.provider_tree [None req-f81b332d-72eb-4149-a6a2-3c8a9c9bec53 tempest-ServersNegativeTestJSON-1008333314 tempest-ServersNegativeTestJSON-1008333314-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 627.696101] env[63088]: INFO nova.compute.manager [-] [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] Took 1.04 seconds to deallocate network for instance. [ 627.699985] env[63088]: DEBUG nova.compute.claims [None req-b1b885bf-cef9-4808-8f16-707ab04499b3 tempest-ServersTestManualDisk-1859582357 tempest-ServersTestManualDisk-1859582357-project-member] [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] Aborting claim: {{(pid=63088) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 627.699985] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b1b885bf-cef9-4808-8f16-707ab04499b3 tempest-ServersTestManualDisk-1859582357 tempest-ServersTestManualDisk-1859582357-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 628.120560] env[63088]: DEBUG nova.compute.manager [None req-0564aade-729d-4cf1-a59d-513384386c54 tempest-AttachInterfacesV270Test-1783457293 tempest-AttachInterfacesV270Test-1783457293-project-member] [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 628.143931] env[63088]: DEBUG nova.scheduler.client.report [None req-f81b332d-72eb-4149-a6a2-3c8a9c9bec53 tempest-ServersNegativeTestJSON-1008333314 tempest-ServersNegativeTestJSON-1008333314-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 628.158100] env[63088]: DEBUG nova.virt.hardware [None req-0564aade-729d-4cf1-a59d-513384386c54 tempest-AttachInterfacesV270Test-1783457293 tempest-AttachInterfacesV270Test-1783457293-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 628.158372] env[63088]: DEBUG nova.virt.hardware [None req-0564aade-729d-4cf1-a59d-513384386c54 tempest-AttachInterfacesV270Test-1783457293 tempest-AttachInterfacesV270Test-1783457293-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 628.158533] env[63088]: DEBUG nova.virt.hardware [None req-0564aade-729d-4cf1-a59d-513384386c54 tempest-AttachInterfacesV270Test-1783457293 tempest-AttachInterfacesV270Test-1783457293-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 628.158717] env[63088]: DEBUG nova.virt.hardware [None req-0564aade-729d-4cf1-a59d-513384386c54 tempest-AttachInterfacesV270Test-1783457293 tempest-AttachInterfacesV270Test-1783457293-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 628.159228] env[63088]: DEBUG nova.virt.hardware [None req-0564aade-729d-4cf1-a59d-513384386c54 tempest-AttachInterfacesV270Test-1783457293 tempest-AttachInterfacesV270Test-1783457293-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 628.159228] env[63088]: DEBUG nova.virt.hardware [None req-0564aade-729d-4cf1-a59d-513384386c54 tempest-AttachInterfacesV270Test-1783457293 tempest-AttachInterfacesV270Test-1783457293-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 628.159409] env[63088]: DEBUG nova.virt.hardware [None req-0564aade-729d-4cf1-a59d-513384386c54 tempest-AttachInterfacesV270Test-1783457293 tempest-AttachInterfacesV270Test-1783457293-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 628.159610] env[63088]: DEBUG nova.virt.hardware [None req-0564aade-729d-4cf1-a59d-513384386c54 tempest-AttachInterfacesV270Test-1783457293 tempest-AttachInterfacesV270Test-1783457293-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 628.159835] env[63088]: DEBUG nova.virt.hardware [None req-0564aade-729d-4cf1-a59d-513384386c54 tempest-AttachInterfacesV270Test-1783457293 tempest-AttachInterfacesV270Test-1783457293-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 628.160049] env[63088]: DEBUG nova.virt.hardware [None req-0564aade-729d-4cf1-a59d-513384386c54 tempest-AttachInterfacesV270Test-1783457293 tempest-AttachInterfacesV270Test-1783457293-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 628.160282] env[63088]: DEBUG nova.virt.hardware [None req-0564aade-729d-4cf1-a59d-513384386c54 tempest-AttachInterfacesV270Test-1783457293 tempest-AttachInterfacesV270Test-1783457293-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 628.161525] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e3faccc-2b24-4c90-9566-7e729b6f37f0 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.171360] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1054dac-874a-4675-8396-6753b1c0951e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.651149] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f81b332d-72eb-4149-a6a2-3c8a9c9bec53 tempest-ServersNegativeTestJSON-1008333314 tempest-ServersNegativeTestJSON-1008333314-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.569s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 628.651149] env[63088]: DEBUG nova.compute.manager [None req-f81b332d-72eb-4149-a6a2-3c8a9c9bec53 tempest-ServersNegativeTestJSON-1008333314 tempest-ServersNegativeTestJSON-1008333314-project-member] [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 628.652371] env[63088]: DEBUG oslo_concurrency.lockutils [None req-79832ead-48a0-4aad-a47e-8e38c2a687da tempest-ServersTestJSON-106593364 tempest-ServersTestJSON-106593364-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 23.449s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 629.035263] env[63088]: DEBUG nova.compute.manager [req-b5f27572-f1e2-4081-9ee4-46769fc2b0b0 req-534127c6-98c0-4f14-82b2-8fde62493f0e service nova] [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] Received event network-changed-c7b197ee-fb06-46ec-9aac-b7fb7874113c {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 629.035503] env[63088]: DEBUG nova.compute.manager [req-b5f27572-f1e2-4081-9ee4-46769fc2b0b0 req-534127c6-98c0-4f14-82b2-8fde62493f0e service nova] [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] Refreshing instance network info cache due to event network-changed-c7b197ee-fb06-46ec-9aac-b7fb7874113c. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 629.035732] env[63088]: DEBUG oslo_concurrency.lockutils [req-b5f27572-f1e2-4081-9ee4-46769fc2b0b0 req-534127c6-98c0-4f14-82b2-8fde62493f0e service nova] Acquiring lock "refresh_cache-421e0f76-024d-4793-8f00-98731bb6b1af" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 629.035873] env[63088]: DEBUG oslo_concurrency.lockutils [req-b5f27572-f1e2-4081-9ee4-46769fc2b0b0 req-534127c6-98c0-4f14-82b2-8fde62493f0e service nova] Acquired lock "refresh_cache-421e0f76-024d-4793-8f00-98731bb6b1af" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 629.036912] env[63088]: DEBUG nova.network.neutron [req-b5f27572-f1e2-4081-9ee4-46769fc2b0b0 req-534127c6-98c0-4f14-82b2-8fde62493f0e service nova] [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] Refreshing network info cache for port c7b197ee-fb06-46ec-9aac-b7fb7874113c {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 629.157429] env[63088]: DEBUG nova.compute.utils [None req-f81b332d-72eb-4149-a6a2-3c8a9c9bec53 tempest-ServersNegativeTestJSON-1008333314 tempest-ServersNegativeTestJSON-1008333314-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 629.167467] env[63088]: DEBUG nova.compute.manager [None req-f81b332d-72eb-4149-a6a2-3c8a9c9bec53 tempest-ServersNegativeTestJSON-1008333314 tempest-ServersNegativeTestJSON-1008333314-project-member] [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 629.167467] env[63088]: DEBUG nova.network.neutron [None req-f81b332d-72eb-4149-a6a2-3c8a9c9bec53 tempest-ServersNegativeTestJSON-1008333314 tempest-ServersNegativeTestJSON-1008333314-project-member] [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 629.224238] env[63088]: DEBUG nova.policy [None req-f81b332d-72eb-4149-a6a2-3c8a9c9bec53 tempest-ServersNegativeTestJSON-1008333314 tempest-ServersNegativeTestJSON-1008333314-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c319878428c249dcb4ac51d8c6fb57ce', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8c3248e405eb41c489836b960823b06f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 629.401170] env[63088]: ERROR nova.compute.manager [None req-0564aade-729d-4cf1-a59d-513384386c54 tempest-AttachInterfacesV270Test-1783457293 tempest-AttachInterfacesV270Test-1783457293-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c7b197ee-fb06-46ec-9aac-b7fb7874113c, please check neutron logs for more information. [ 629.401170] env[63088]: ERROR nova.compute.manager Traceback (most recent call last): [ 629.401170] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 629.401170] env[63088]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 629.401170] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 629.401170] env[63088]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 629.401170] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 629.401170] env[63088]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 629.401170] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 629.401170] env[63088]: ERROR nova.compute.manager self.force_reraise() [ 629.401170] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 629.401170] env[63088]: ERROR nova.compute.manager raise self.value [ 629.401170] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 629.401170] env[63088]: ERROR nova.compute.manager updated_port = self._update_port( [ 629.401170] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 629.401170] env[63088]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 629.401715] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 629.401715] env[63088]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 629.401715] env[63088]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c7b197ee-fb06-46ec-9aac-b7fb7874113c, please check neutron logs for more information. [ 629.401715] env[63088]: ERROR nova.compute.manager [ 629.401715] env[63088]: Traceback (most recent call last): [ 629.401715] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 629.401715] env[63088]: listener.cb(fileno) [ 629.401715] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 629.401715] env[63088]: result = function(*args, **kwargs) [ 629.401715] env[63088]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 629.401715] env[63088]: return func(*args, **kwargs) [ 629.401715] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 629.401715] env[63088]: raise e [ 629.401715] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 629.401715] env[63088]: nwinfo = self.network_api.allocate_for_instance( [ 629.401715] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 629.401715] env[63088]: created_port_ids = self._update_ports_for_instance( [ 629.401715] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 629.401715] env[63088]: with excutils.save_and_reraise_exception(): [ 629.401715] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 629.401715] env[63088]: self.force_reraise() [ 629.401715] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 629.401715] env[63088]: raise self.value [ 629.401715] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 629.401715] env[63088]: updated_port = self._update_port( [ 629.401715] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 629.401715] env[63088]: _ensure_no_port_binding_failure(port) [ 629.401715] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 629.401715] env[63088]: raise exception.PortBindingFailed(port_id=port['id']) [ 629.402374] env[63088]: nova.exception.PortBindingFailed: Binding failed for port c7b197ee-fb06-46ec-9aac-b7fb7874113c, please check neutron logs for more information. [ 629.402374] env[63088]: Removing descriptor: 15 [ 629.402374] env[63088]: ERROR nova.compute.manager [None req-0564aade-729d-4cf1-a59d-513384386c54 tempest-AttachInterfacesV270Test-1783457293 tempest-AttachInterfacesV270Test-1783457293-project-member] [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c7b197ee-fb06-46ec-9aac-b7fb7874113c, please check neutron logs for more information. [ 629.402374] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] Traceback (most recent call last): [ 629.402374] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 629.402374] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] yield resources [ 629.402374] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 629.402374] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] self.driver.spawn(context, instance, image_meta, [ 629.402374] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 629.402374] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] self._vmops.spawn(context, instance, image_meta, injected_files, [ 629.402374] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 629.402374] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] vm_ref = self.build_virtual_machine(instance, [ 629.402673] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 629.402673] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] vif_infos = vmwarevif.get_vif_info(self._session, [ 629.402673] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 629.402673] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] for vif in network_info: [ 629.402673] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 629.402673] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] return self._sync_wrapper(fn, *args, **kwargs) [ 629.402673] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 629.402673] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] self.wait() [ 629.402673] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 629.402673] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] self[:] = self._gt.wait() [ 629.402673] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 629.402673] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] return self._exit_event.wait() [ 629.402673] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 629.402956] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] result = hub.switch() [ 629.402956] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 629.402956] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] return self.greenlet.switch() [ 629.402956] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 629.402956] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] result = function(*args, **kwargs) [ 629.402956] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 629.402956] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] return func(*args, **kwargs) [ 629.402956] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 629.402956] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] raise e [ 629.402956] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 629.402956] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] nwinfo = self.network_api.allocate_for_instance( [ 629.402956] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 629.402956] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] created_port_ids = self._update_ports_for_instance( [ 629.403320] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 629.403320] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] with excutils.save_and_reraise_exception(): [ 629.403320] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 629.403320] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] self.force_reraise() [ 629.403320] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 629.403320] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] raise self.value [ 629.403320] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 629.403320] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] updated_port = self._update_port( [ 629.403320] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 629.403320] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] _ensure_no_port_binding_failure(port) [ 629.403320] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 629.403320] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] raise exception.PortBindingFailed(port_id=port['id']) [ 629.403591] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] nova.exception.PortBindingFailed: Binding failed for port c7b197ee-fb06-46ec-9aac-b7fb7874113c, please check neutron logs for more information. [ 629.403591] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] [ 629.403591] env[63088]: INFO nova.compute.manager [None req-0564aade-729d-4cf1-a59d-513384386c54 tempest-AttachInterfacesV270Test-1783457293 tempest-AttachInterfacesV270Test-1783457293-project-member] [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] Terminating instance [ 629.404751] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0564aade-729d-4cf1-a59d-513384386c54 tempest-AttachInterfacesV270Test-1783457293 tempest-AttachInterfacesV270Test-1783457293-project-member] Acquiring lock "refresh_cache-421e0f76-024d-4793-8f00-98731bb6b1af" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 629.562156] env[63088]: DEBUG nova.network.neutron [req-b5f27572-f1e2-4081-9ee4-46769fc2b0b0 req-534127c6-98c0-4f14-82b2-8fde62493f0e service nova] [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 629.579523] env[63088]: DEBUG nova.network.neutron [None req-f81b332d-72eb-4149-a6a2-3c8a9c9bec53 tempest-ServersNegativeTestJSON-1008333314 tempest-ServersNegativeTestJSON-1008333314-project-member] [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] Successfully created port: a0fc3b2d-a2f9-457b-acc2-16a9ac49a5bd {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 629.659919] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a622c7b-ebe9-40f2-a931-876d809ee1cd {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.668755] env[63088]: DEBUG nova.compute.manager [None req-f81b332d-72eb-4149-a6a2-3c8a9c9bec53 tempest-ServersNegativeTestJSON-1008333314 tempest-ServersNegativeTestJSON-1008333314-project-member] [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 629.672022] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fca4471c-3b01-4d79-9593-ede8afd40f65 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.711862] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7cee552-4d8c-4d9f-80e4-5fb43f92aeba {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.718813] env[63088]: DEBUG nova.network.neutron [req-b5f27572-f1e2-4081-9ee4-46769fc2b0b0 req-534127c6-98c0-4f14-82b2-8fde62493f0e service nova] [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 629.728417] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20bf1cf8-6108-4ca7-b6ee-bbed2f9343f1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.749089] env[63088]: DEBUG nova.compute.provider_tree [None req-79832ead-48a0-4aad-a47e-8e38c2a687da tempest-ServersTestJSON-106593364 tempest-ServersTestJSON-106593364-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 630.221591] env[63088]: DEBUG oslo_concurrency.lockutils [req-b5f27572-f1e2-4081-9ee4-46769fc2b0b0 req-534127c6-98c0-4f14-82b2-8fde62493f0e service nova] Releasing lock "refresh_cache-421e0f76-024d-4793-8f00-98731bb6b1af" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 630.222055] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0564aade-729d-4cf1-a59d-513384386c54 tempest-AttachInterfacesV270Test-1783457293 tempest-AttachInterfacesV270Test-1783457293-project-member] Acquired lock "refresh_cache-421e0f76-024d-4793-8f00-98731bb6b1af" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 630.222240] env[63088]: DEBUG nova.network.neutron [None req-0564aade-729d-4cf1-a59d-513384386c54 tempest-AttachInterfacesV270Test-1783457293 tempest-AttachInterfacesV270Test-1783457293-project-member] [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 630.255018] env[63088]: DEBUG nova.scheduler.client.report [None req-79832ead-48a0-4aad-a47e-8e38c2a687da tempest-ServersTestJSON-106593364 tempest-ServersTestJSON-106593364-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 630.673097] env[63088]: DEBUG nova.compute.manager [req-0f4a2afd-579d-47e7-af1a-c8d027868e2a req-9effe0ed-136e-4ee1-9e8c-7fd7d273b236 service nova] [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] Received event network-changed-a0fc3b2d-a2f9-457b-acc2-16a9ac49a5bd {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 630.673283] env[63088]: DEBUG nova.compute.manager [req-0f4a2afd-579d-47e7-af1a-c8d027868e2a req-9effe0ed-136e-4ee1-9e8c-7fd7d273b236 service nova] [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] Refreshing instance network info cache due to event network-changed-a0fc3b2d-a2f9-457b-acc2-16a9ac49a5bd. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 630.673493] env[63088]: DEBUG oslo_concurrency.lockutils [req-0f4a2afd-579d-47e7-af1a-c8d027868e2a req-9effe0ed-136e-4ee1-9e8c-7fd7d273b236 service nova] Acquiring lock "refresh_cache-5b778247-1434-447e-8ff1-7359bb9e8ea2" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 630.673631] env[63088]: DEBUG oslo_concurrency.lockutils [req-0f4a2afd-579d-47e7-af1a-c8d027868e2a req-9effe0ed-136e-4ee1-9e8c-7fd7d273b236 service nova] Acquired lock "refresh_cache-5b778247-1434-447e-8ff1-7359bb9e8ea2" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 630.673790] env[63088]: DEBUG nova.network.neutron [req-0f4a2afd-579d-47e7-af1a-c8d027868e2a req-9effe0ed-136e-4ee1-9e8c-7fd7d273b236 service nova] [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] Refreshing network info cache for port a0fc3b2d-a2f9-457b-acc2-16a9ac49a5bd {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 630.685942] env[63088]: DEBUG nova.compute.manager [None req-f81b332d-72eb-4149-a6a2-3c8a9c9bec53 tempest-ServersNegativeTestJSON-1008333314 tempest-ServersNegativeTestJSON-1008333314-project-member] [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 630.721604] env[63088]: DEBUG nova.virt.hardware [None req-f81b332d-72eb-4149-a6a2-3c8a9c9bec53 tempest-ServersNegativeTestJSON-1008333314 tempest-ServersNegativeTestJSON-1008333314-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 630.721933] env[63088]: DEBUG nova.virt.hardware [None req-f81b332d-72eb-4149-a6a2-3c8a9c9bec53 tempest-ServersNegativeTestJSON-1008333314 tempest-ServersNegativeTestJSON-1008333314-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 630.722113] env[63088]: DEBUG nova.virt.hardware [None req-f81b332d-72eb-4149-a6a2-3c8a9c9bec53 tempest-ServersNegativeTestJSON-1008333314 tempest-ServersNegativeTestJSON-1008333314-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 630.722302] env[63088]: DEBUG nova.virt.hardware [None req-f81b332d-72eb-4149-a6a2-3c8a9c9bec53 tempest-ServersNegativeTestJSON-1008333314 tempest-ServersNegativeTestJSON-1008333314-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 630.722449] env[63088]: DEBUG nova.virt.hardware [None req-f81b332d-72eb-4149-a6a2-3c8a9c9bec53 tempest-ServersNegativeTestJSON-1008333314 tempest-ServersNegativeTestJSON-1008333314-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 630.722594] env[63088]: DEBUG nova.virt.hardware [None req-f81b332d-72eb-4149-a6a2-3c8a9c9bec53 tempest-ServersNegativeTestJSON-1008333314 tempest-ServersNegativeTestJSON-1008333314-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 630.722795] env[63088]: DEBUG nova.virt.hardware [None req-f81b332d-72eb-4149-a6a2-3c8a9c9bec53 tempest-ServersNegativeTestJSON-1008333314 tempest-ServersNegativeTestJSON-1008333314-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 630.722951] env[63088]: DEBUG nova.virt.hardware [None req-f81b332d-72eb-4149-a6a2-3c8a9c9bec53 tempest-ServersNegativeTestJSON-1008333314 tempest-ServersNegativeTestJSON-1008333314-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 630.723159] env[63088]: DEBUG nova.virt.hardware [None req-f81b332d-72eb-4149-a6a2-3c8a9c9bec53 tempest-ServersNegativeTestJSON-1008333314 tempest-ServersNegativeTestJSON-1008333314-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 630.723615] env[63088]: DEBUG nova.virt.hardware [None req-f81b332d-72eb-4149-a6a2-3c8a9c9bec53 tempest-ServersNegativeTestJSON-1008333314 tempest-ServersNegativeTestJSON-1008333314-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 630.723791] env[63088]: DEBUG nova.virt.hardware [None req-f81b332d-72eb-4149-a6a2-3c8a9c9bec53 tempest-ServersNegativeTestJSON-1008333314 tempest-ServersNegativeTestJSON-1008333314-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 630.725243] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f202488-59d0-4fb1-81ca-b86b047b7021 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.738293] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b49d05dc-283e-4543-a075-b3461cd7b14c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.757613] env[63088]: DEBUG nova.network.neutron [None req-0564aade-729d-4cf1-a59d-513384386c54 tempest-AttachInterfacesV270Test-1783457293 tempest-AttachInterfacesV270Test-1783457293-project-member] [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 630.763031] env[63088]: DEBUG oslo_concurrency.lockutils [None req-79832ead-48a0-4aad-a47e-8e38c2a687da tempest-ServersTestJSON-106593364 tempest-ServersTestJSON-106593364-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.110s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 630.763031] env[63088]: ERROR nova.compute.manager [None req-79832ead-48a0-4aad-a47e-8e38c2a687da tempest-ServersTestJSON-106593364 tempest-ServersTestJSON-106593364-project-member] [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 203eaeff-191a-4386-871b-58f5e69aa834, please check neutron logs for more information. [ 630.763031] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] Traceback (most recent call last): [ 630.763031] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 630.763031] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] self.driver.spawn(context, instance, image_meta, [ 630.763031] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 630.763031] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] self._vmops.spawn(context, instance, image_meta, injected_files, [ 630.763031] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 630.763031] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] vm_ref = self.build_virtual_machine(instance, [ 630.763361] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 630.763361] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] vif_infos = vmwarevif.get_vif_info(self._session, [ 630.763361] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 630.763361] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] for vif in network_info: [ 630.763361] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 630.763361] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] return self._sync_wrapper(fn, *args, **kwargs) [ 630.763361] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 630.763361] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] self.wait() [ 630.763361] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 630.763361] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] self[:] = self._gt.wait() [ 630.763361] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 630.763361] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] return self._exit_event.wait() [ 630.763361] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 630.763659] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] result = hub.switch() [ 630.763659] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 630.763659] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] return self.greenlet.switch() [ 630.763659] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 630.763659] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] result = function(*args, **kwargs) [ 630.763659] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 630.763659] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] return func(*args, **kwargs) [ 630.763659] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 630.763659] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] raise e [ 630.763659] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 630.763659] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] nwinfo = self.network_api.allocate_for_instance( [ 630.763659] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 630.763659] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] created_port_ids = self._update_ports_for_instance( [ 630.764030] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 630.764030] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] with excutils.save_and_reraise_exception(): [ 630.764030] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 630.764030] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] self.force_reraise() [ 630.764030] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 630.764030] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] raise self.value [ 630.764030] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 630.764030] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] updated_port = self._update_port( [ 630.764030] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 630.764030] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] _ensure_no_port_binding_failure(port) [ 630.764030] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 630.764030] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] raise exception.PortBindingFailed(port_id=port['id']) [ 630.764322] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] nova.exception.PortBindingFailed: Binding failed for port 203eaeff-191a-4386-871b-58f5e69aa834, please check neutron logs for more information. [ 630.764322] env[63088]: ERROR nova.compute.manager [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] [ 630.764322] env[63088]: DEBUG nova.compute.utils [None req-79832ead-48a0-4aad-a47e-8e38c2a687da tempest-ServersTestJSON-106593364 tempest-ServersTestJSON-106593364-project-member] [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] Binding failed for port 203eaeff-191a-4386-871b-58f5e69aa834, please check neutron logs for more information. {{(pid=63088) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 630.766061] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a9773ab2-f8b9-4102-8e41-5ee881a8637e tempest-FloatingIPsAssociationNegativeTestJSON-189613961 tempest-FloatingIPsAssociationNegativeTestJSON-189613961-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 21.063s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 630.773021] env[63088]: DEBUG nova.compute.manager [None req-79832ead-48a0-4aad-a47e-8e38c2a687da tempest-ServersTestJSON-106593364 tempest-ServersTestJSON-106593364-project-member] [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] Build of instance 34c01c2c-519b-44a1-aba2-c26e2c0b4737 was re-scheduled: Binding failed for port 203eaeff-191a-4386-871b-58f5e69aa834, please check neutron logs for more information. {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 630.773021] env[63088]: DEBUG nova.compute.manager [None req-79832ead-48a0-4aad-a47e-8e38c2a687da tempest-ServersTestJSON-106593364 tempest-ServersTestJSON-106593364-project-member] [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] Unplugging VIFs for instance {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 630.773021] env[63088]: DEBUG oslo_concurrency.lockutils [None req-79832ead-48a0-4aad-a47e-8e38c2a687da tempest-ServersTestJSON-106593364 tempest-ServersTestJSON-106593364-project-member] Acquiring lock "refresh_cache-34c01c2c-519b-44a1-aba2-c26e2c0b4737" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 630.773021] env[63088]: DEBUG oslo_concurrency.lockutils [None req-79832ead-48a0-4aad-a47e-8e38c2a687da tempest-ServersTestJSON-106593364 tempest-ServersTestJSON-106593364-project-member] Acquired lock "refresh_cache-34c01c2c-519b-44a1-aba2-c26e2c0b4737" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 630.773021] env[63088]: DEBUG nova.network.neutron [None req-79832ead-48a0-4aad-a47e-8e38c2a687da tempest-ServersTestJSON-106593364 tempest-ServersTestJSON-106593364-project-member] [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 630.917022] env[63088]: DEBUG nova.network.neutron [None req-0564aade-729d-4cf1-a59d-513384386c54 tempest-AttachInterfacesV270Test-1783457293 tempest-AttachInterfacesV270Test-1783457293-project-member] [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 631.070504] env[63088]: DEBUG nova.compute.manager [req-dc356ae1-1ac1-401d-9abe-6fecb8c56b0f req-8704a3ee-2406-4a88-98a0-52e6e7ba4bf0 service nova] [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] Received event network-vif-deleted-c7b197ee-fb06-46ec-9aac-b7fb7874113c {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 631.093614] env[63088]: ERROR nova.compute.manager [None req-f81b332d-72eb-4149-a6a2-3c8a9c9bec53 tempest-ServersNegativeTestJSON-1008333314 tempest-ServersNegativeTestJSON-1008333314-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a0fc3b2d-a2f9-457b-acc2-16a9ac49a5bd, please check neutron logs for more information. [ 631.093614] env[63088]: ERROR nova.compute.manager Traceback (most recent call last): [ 631.093614] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 631.093614] env[63088]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 631.093614] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 631.093614] env[63088]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 631.093614] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 631.093614] env[63088]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 631.093614] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 631.093614] env[63088]: ERROR nova.compute.manager self.force_reraise() [ 631.093614] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 631.093614] env[63088]: ERROR nova.compute.manager raise self.value [ 631.093614] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 631.093614] env[63088]: ERROR nova.compute.manager updated_port = self._update_port( [ 631.093614] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 631.093614] env[63088]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 631.094066] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 631.094066] env[63088]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 631.094066] env[63088]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a0fc3b2d-a2f9-457b-acc2-16a9ac49a5bd, please check neutron logs for more information. [ 631.094066] env[63088]: ERROR nova.compute.manager [ 631.094066] env[63088]: Traceback (most recent call last): [ 631.094066] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 631.094066] env[63088]: listener.cb(fileno) [ 631.094066] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 631.094066] env[63088]: result = function(*args, **kwargs) [ 631.094066] env[63088]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 631.094066] env[63088]: return func(*args, **kwargs) [ 631.094066] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 631.094066] env[63088]: raise e [ 631.094066] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 631.094066] env[63088]: nwinfo = self.network_api.allocate_for_instance( [ 631.094066] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 631.094066] env[63088]: created_port_ids = self._update_ports_for_instance( [ 631.094066] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 631.094066] env[63088]: with excutils.save_and_reraise_exception(): [ 631.094066] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 631.094066] env[63088]: self.force_reraise() [ 631.094066] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 631.094066] env[63088]: raise self.value [ 631.094066] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 631.094066] env[63088]: updated_port = self._update_port( [ 631.094066] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 631.094066] env[63088]: _ensure_no_port_binding_failure(port) [ 631.094066] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 631.094066] env[63088]: raise exception.PortBindingFailed(port_id=port['id']) [ 631.094801] env[63088]: nova.exception.PortBindingFailed: Binding failed for port a0fc3b2d-a2f9-457b-acc2-16a9ac49a5bd, please check neutron logs for more information. [ 631.094801] env[63088]: Removing descriptor: 17 [ 631.094801] env[63088]: ERROR nova.compute.manager [None req-f81b332d-72eb-4149-a6a2-3c8a9c9bec53 tempest-ServersNegativeTestJSON-1008333314 tempest-ServersNegativeTestJSON-1008333314-project-member] [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a0fc3b2d-a2f9-457b-acc2-16a9ac49a5bd, please check neutron logs for more information. [ 631.094801] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] Traceback (most recent call last): [ 631.094801] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 631.094801] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] yield resources [ 631.094801] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 631.094801] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] self.driver.spawn(context, instance, image_meta, [ 631.094801] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 631.094801] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 631.094801] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 631.094801] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] vm_ref = self.build_virtual_machine(instance, [ 631.095132] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 631.095132] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] vif_infos = vmwarevif.get_vif_info(self._session, [ 631.095132] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 631.095132] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] for vif in network_info: [ 631.095132] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 631.095132] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] return self._sync_wrapper(fn, *args, **kwargs) [ 631.095132] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 631.095132] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] self.wait() [ 631.095132] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 631.095132] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] self[:] = self._gt.wait() [ 631.095132] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 631.095132] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] return self._exit_event.wait() [ 631.095132] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 631.095538] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] result = hub.switch() [ 631.095538] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 631.095538] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] return self.greenlet.switch() [ 631.095538] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 631.095538] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] result = function(*args, **kwargs) [ 631.095538] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 631.095538] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] return func(*args, **kwargs) [ 631.095538] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 631.095538] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] raise e [ 631.095538] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 631.095538] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] nwinfo = self.network_api.allocate_for_instance( [ 631.095538] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 631.095538] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] created_port_ids = self._update_ports_for_instance( [ 631.095870] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 631.095870] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] with excutils.save_and_reraise_exception(): [ 631.095870] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 631.095870] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] self.force_reraise() [ 631.095870] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 631.095870] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] raise self.value [ 631.095870] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 631.095870] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] updated_port = self._update_port( [ 631.095870] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 631.095870] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] _ensure_no_port_binding_failure(port) [ 631.095870] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 631.095870] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] raise exception.PortBindingFailed(port_id=port['id']) [ 631.096189] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] nova.exception.PortBindingFailed: Binding failed for port a0fc3b2d-a2f9-457b-acc2-16a9ac49a5bd, please check neutron logs for more information. [ 631.096189] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] [ 631.096189] env[63088]: INFO nova.compute.manager [None req-f81b332d-72eb-4149-a6a2-3c8a9c9bec53 tempest-ServersNegativeTestJSON-1008333314 tempest-ServersNegativeTestJSON-1008333314-project-member] [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] Terminating instance [ 631.096839] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f81b332d-72eb-4149-a6a2-3c8a9c9bec53 tempest-ServersNegativeTestJSON-1008333314 tempest-ServersNegativeTestJSON-1008333314-project-member] Acquiring lock "refresh_cache-5b778247-1434-447e-8ff1-7359bb9e8ea2" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 631.193985] env[63088]: DEBUG nova.network.neutron [req-0f4a2afd-579d-47e7-af1a-c8d027868e2a req-9effe0ed-136e-4ee1-9e8c-7fd7d273b236 service nova] [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 631.298220] env[63088]: DEBUG nova.network.neutron [None req-79832ead-48a0-4aad-a47e-8e38c2a687da tempest-ServersTestJSON-106593364 tempest-ServersTestJSON-106593364-project-member] [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 631.353999] env[63088]: DEBUG nova.network.neutron [req-0f4a2afd-579d-47e7-af1a-c8d027868e2a req-9effe0ed-136e-4ee1-9e8c-7fd7d273b236 service nova] [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 631.419916] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0564aade-729d-4cf1-a59d-513384386c54 tempest-AttachInterfacesV270Test-1783457293 tempest-AttachInterfacesV270Test-1783457293-project-member] Releasing lock "refresh_cache-421e0f76-024d-4793-8f00-98731bb6b1af" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 631.420342] env[63088]: DEBUG nova.compute.manager [None req-0564aade-729d-4cf1-a59d-513384386c54 tempest-AttachInterfacesV270Test-1783457293 tempest-AttachInterfacesV270Test-1783457293-project-member] [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 631.420532] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-0564aade-729d-4cf1-a59d-513384386c54 tempest-AttachInterfacesV270Test-1783457293 tempest-AttachInterfacesV270Test-1783457293-project-member] [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 631.420839] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c0eb6023-dc6f-4496-ba5c-31255b67e802 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.433083] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e88d6b0-6ea4-42ae-8c05-1ce729ae187d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.449088] env[63088]: DEBUG nova.network.neutron [None req-79832ead-48a0-4aad-a47e-8e38c2a687da tempest-ServersTestJSON-106593364 tempest-ServersTestJSON-106593364-project-member] [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 631.459746] env[63088]: WARNING nova.virt.vmwareapi.vmops [None req-0564aade-729d-4cf1-a59d-513384386c54 tempest-AttachInterfacesV270Test-1783457293 tempest-AttachInterfacesV270Test-1783457293-project-member] [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 421e0f76-024d-4793-8f00-98731bb6b1af could not be found. [ 631.459968] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-0564aade-729d-4cf1-a59d-513384386c54 tempest-AttachInterfacesV270Test-1783457293 tempest-AttachInterfacesV270Test-1783457293-project-member] [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 631.460161] env[63088]: INFO nova.compute.manager [None req-0564aade-729d-4cf1-a59d-513384386c54 tempest-AttachInterfacesV270Test-1783457293 tempest-AttachInterfacesV270Test-1783457293-project-member] [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] Took 0.04 seconds to destroy the instance on the hypervisor. [ 631.460401] env[63088]: DEBUG oslo.service.loopingcall [None req-0564aade-729d-4cf1-a59d-513384386c54 tempest-AttachInterfacesV270Test-1783457293 tempest-AttachInterfacesV270Test-1783457293-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 631.463705] env[63088]: DEBUG nova.compute.manager [-] [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 631.463807] env[63088]: DEBUG nova.network.neutron [-] [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 631.493172] env[63088]: DEBUG nova.network.neutron [-] [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 631.775679] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94c57f4b-f0d9-463e-9aa1-636cc6cf8f1a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.785795] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dc8a067-c64c-4524-b754-ea7fbccea867 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.827166] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbc73c5e-4a24-4369-aaec-871a69498ab3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.836979] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b53cd050-a128-4320-bd9f-284b2465dbf6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.855762] env[63088]: DEBUG nova.compute.provider_tree [None req-a9773ab2-f8b9-4102-8e41-5ee881a8637e tempest-FloatingIPsAssociationNegativeTestJSON-189613961 tempest-FloatingIPsAssociationNegativeTestJSON-189613961-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 631.860412] env[63088]: DEBUG oslo_concurrency.lockutils [req-0f4a2afd-579d-47e7-af1a-c8d027868e2a req-9effe0ed-136e-4ee1-9e8c-7fd7d273b236 service nova] Releasing lock "refresh_cache-5b778247-1434-447e-8ff1-7359bb9e8ea2" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 631.861317] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f81b332d-72eb-4149-a6a2-3c8a9c9bec53 tempest-ServersNegativeTestJSON-1008333314 tempest-ServersNegativeTestJSON-1008333314-project-member] Acquired lock "refresh_cache-5b778247-1434-447e-8ff1-7359bb9e8ea2" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 631.862470] env[63088]: DEBUG nova.network.neutron [None req-f81b332d-72eb-4149-a6a2-3c8a9c9bec53 tempest-ServersNegativeTestJSON-1008333314 tempest-ServersNegativeTestJSON-1008333314-project-member] [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 631.953788] env[63088]: DEBUG oslo_concurrency.lockutils [None req-79832ead-48a0-4aad-a47e-8e38c2a687da tempest-ServersTestJSON-106593364 tempest-ServersTestJSON-106593364-project-member] Releasing lock "refresh_cache-34c01c2c-519b-44a1-aba2-c26e2c0b4737" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 631.953788] env[63088]: DEBUG nova.compute.manager [None req-79832ead-48a0-4aad-a47e-8e38c2a687da tempest-ServersTestJSON-106593364 tempest-ServersTestJSON-106593364-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 631.953788] env[63088]: DEBUG nova.compute.manager [None req-79832ead-48a0-4aad-a47e-8e38c2a687da tempest-ServersTestJSON-106593364 tempest-ServersTestJSON-106593364-project-member] [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 631.953788] env[63088]: DEBUG nova.network.neutron [None req-79832ead-48a0-4aad-a47e-8e38c2a687da tempest-ServersTestJSON-106593364 tempest-ServersTestJSON-106593364-project-member] [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 631.998454] env[63088]: DEBUG nova.network.neutron [-] [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 632.126688] env[63088]: DEBUG nova.network.neutron [None req-79832ead-48a0-4aad-a47e-8e38c2a687da tempest-ServersTestJSON-106593364 tempest-ServersTestJSON-106593364-project-member] [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 632.359491] env[63088]: DEBUG nova.scheduler.client.report [None req-a9773ab2-f8b9-4102-8e41-5ee881a8637e tempest-FloatingIPsAssociationNegativeTestJSON-189613961 tempest-FloatingIPsAssociationNegativeTestJSON-189613961-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 632.391511] env[63088]: DEBUG nova.network.neutron [None req-f81b332d-72eb-4149-a6a2-3c8a9c9bec53 tempest-ServersNegativeTestJSON-1008333314 tempest-ServersNegativeTestJSON-1008333314-project-member] [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 632.503844] env[63088]: INFO nova.compute.manager [-] [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] Took 1.04 seconds to deallocate network for instance. [ 632.507171] env[63088]: DEBUG nova.compute.claims [None req-0564aade-729d-4cf1-a59d-513384386c54 tempest-AttachInterfacesV270Test-1783457293 tempest-AttachInterfacesV270Test-1783457293-project-member] [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] Aborting claim: {{(pid=63088) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 632.507171] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0564aade-729d-4cf1-a59d-513384386c54 tempest-AttachInterfacesV270Test-1783457293 tempest-AttachInterfacesV270Test-1783457293-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 632.530569] env[63088]: DEBUG nova.network.neutron [None req-f81b332d-72eb-4149-a6a2-3c8a9c9bec53 tempest-ServersNegativeTestJSON-1008333314 tempest-ServersNegativeTestJSON-1008333314-project-member] [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 632.629715] env[63088]: DEBUG nova.network.neutron [None req-79832ead-48a0-4aad-a47e-8e38c2a687da tempest-ServersTestJSON-106593364 tempest-ServersTestJSON-106593364-project-member] [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 632.756358] env[63088]: DEBUG nova.compute.manager [req-9447d532-18a6-4a2b-a585-e05a03d22eac req-bb4bfa91-b433-45af-bb48-1734144e1dce service nova] [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] Received event network-vif-deleted-a0fc3b2d-a2f9-457b-acc2-16a9ac49a5bd {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 632.868123] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a9773ab2-f8b9-4102-8e41-5ee881a8637e tempest-FloatingIPsAssociationNegativeTestJSON-189613961 tempest-FloatingIPsAssociationNegativeTestJSON-189613961-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.102s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 632.869300] env[63088]: ERROR nova.compute.manager [None req-a9773ab2-f8b9-4102-8e41-5ee881a8637e tempest-FloatingIPsAssociationNegativeTestJSON-189613961 tempest-FloatingIPsAssociationNegativeTestJSON-189613961-project-member] [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8d873b9d-f3a1-455e-aeb3-c68d1fa4fea7, please check neutron logs for more information. [ 632.869300] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] Traceback (most recent call last): [ 632.869300] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 632.869300] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] self.driver.spawn(context, instance, image_meta, [ 632.869300] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 632.869300] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 632.869300] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 632.869300] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] vm_ref = self.build_virtual_machine(instance, [ 632.869300] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 632.869300] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] vif_infos = vmwarevif.get_vif_info(self._session, [ 632.869300] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 632.869663] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] for vif in network_info: [ 632.869663] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 632.869663] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] return self._sync_wrapper(fn, *args, **kwargs) [ 632.869663] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 632.869663] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] self.wait() [ 632.869663] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 632.869663] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] self[:] = self._gt.wait() [ 632.869663] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 632.869663] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] return self._exit_event.wait() [ 632.869663] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 632.869663] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] result = hub.switch() [ 632.869663] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 632.869663] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] return self.greenlet.switch() [ 632.869949] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 632.869949] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] result = function(*args, **kwargs) [ 632.869949] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 632.869949] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] return func(*args, **kwargs) [ 632.869949] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 632.869949] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] raise e [ 632.869949] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 632.869949] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] nwinfo = self.network_api.allocate_for_instance( [ 632.869949] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 632.869949] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] created_port_ids = self._update_ports_for_instance( [ 632.869949] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 632.869949] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] with excutils.save_and_reraise_exception(): [ 632.869949] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 632.870321] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] self.force_reraise() [ 632.870321] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 632.870321] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] raise self.value [ 632.870321] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 632.870321] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] updated_port = self._update_port( [ 632.870321] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 632.870321] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] _ensure_no_port_binding_failure(port) [ 632.870321] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 632.870321] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] raise exception.PortBindingFailed(port_id=port['id']) [ 632.870321] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] nova.exception.PortBindingFailed: Binding failed for port 8d873b9d-f3a1-455e-aeb3-c68d1fa4fea7, please check neutron logs for more information. [ 632.870321] env[63088]: ERROR nova.compute.manager [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] [ 632.870589] env[63088]: DEBUG nova.compute.utils [None req-a9773ab2-f8b9-4102-8e41-5ee881a8637e tempest-FloatingIPsAssociationNegativeTestJSON-189613961 tempest-FloatingIPsAssociationNegativeTestJSON-189613961-project-member] [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] Binding failed for port 8d873b9d-f3a1-455e-aeb3-c68d1fa4fea7, please check neutron logs for more information. {{(pid=63088) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 632.870874] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bd71c985-753f-4eb9-b297-7d458ea844aa tempest-ImagesOneServerNegativeTestJSON-1975365234 tempest-ImagesOneServerNegativeTestJSON-1975365234-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 22.446s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 632.874480] env[63088]: DEBUG nova.compute.manager [None req-a9773ab2-f8b9-4102-8e41-5ee881a8637e tempest-FloatingIPsAssociationNegativeTestJSON-189613961 tempest-FloatingIPsAssociationNegativeTestJSON-189613961-project-member] [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] Build of instance 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4 was re-scheduled: Binding failed for port 8d873b9d-f3a1-455e-aeb3-c68d1fa4fea7, please check neutron logs for more information. {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 632.875609] env[63088]: DEBUG nova.compute.manager [None req-a9773ab2-f8b9-4102-8e41-5ee881a8637e tempest-FloatingIPsAssociationNegativeTestJSON-189613961 tempest-FloatingIPsAssociationNegativeTestJSON-189613961-project-member] [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] Unplugging VIFs for instance {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 632.875795] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a9773ab2-f8b9-4102-8e41-5ee881a8637e tempest-FloatingIPsAssociationNegativeTestJSON-189613961 tempest-FloatingIPsAssociationNegativeTestJSON-189613961-project-member] Acquiring lock "refresh_cache-9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 632.876010] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a9773ab2-f8b9-4102-8e41-5ee881a8637e tempest-FloatingIPsAssociationNegativeTestJSON-189613961 tempest-FloatingIPsAssociationNegativeTestJSON-189613961-project-member] Acquired lock "refresh_cache-9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 632.876190] env[63088]: DEBUG nova.network.neutron [None req-a9773ab2-f8b9-4102-8e41-5ee881a8637e tempest-FloatingIPsAssociationNegativeTestJSON-189613961 tempest-FloatingIPsAssociationNegativeTestJSON-189613961-project-member] [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 633.037505] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f81b332d-72eb-4149-a6a2-3c8a9c9bec53 tempest-ServersNegativeTestJSON-1008333314 tempest-ServersNegativeTestJSON-1008333314-project-member] Releasing lock "refresh_cache-5b778247-1434-447e-8ff1-7359bb9e8ea2" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 633.037998] env[63088]: DEBUG nova.compute.manager [None req-f81b332d-72eb-4149-a6a2-3c8a9c9bec53 tempest-ServersNegativeTestJSON-1008333314 tempest-ServersNegativeTestJSON-1008333314-project-member] [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 633.038212] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-f81b332d-72eb-4149-a6a2-3c8a9c9bec53 tempest-ServersNegativeTestJSON-1008333314 tempest-ServersNegativeTestJSON-1008333314-project-member] [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 633.038532] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-06daa388-5918-451d-bc8a-e8b6e4186a61 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.049687] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec9fd146-ac33-4638-bf15-54cfc55823f7 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.074736] env[63088]: WARNING nova.virt.vmwareapi.vmops [None req-f81b332d-72eb-4149-a6a2-3c8a9c9bec53 tempest-ServersNegativeTestJSON-1008333314 tempest-ServersNegativeTestJSON-1008333314-project-member] [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5b778247-1434-447e-8ff1-7359bb9e8ea2 could not be found. [ 633.074903] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-f81b332d-72eb-4149-a6a2-3c8a9c9bec53 tempest-ServersNegativeTestJSON-1008333314 tempest-ServersNegativeTestJSON-1008333314-project-member] [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 633.075094] env[63088]: INFO nova.compute.manager [None req-f81b332d-72eb-4149-a6a2-3c8a9c9bec53 tempest-ServersNegativeTestJSON-1008333314 tempest-ServersNegativeTestJSON-1008333314-project-member] [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] Took 0.04 seconds to destroy the instance on the hypervisor. [ 633.076295] env[63088]: DEBUG oslo.service.loopingcall [None req-f81b332d-72eb-4149-a6a2-3c8a9c9bec53 tempest-ServersNegativeTestJSON-1008333314 tempest-ServersNegativeTestJSON-1008333314-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 633.076295] env[63088]: DEBUG nova.compute.manager [-] [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 633.076295] env[63088]: DEBUG nova.network.neutron [-] [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 633.092821] env[63088]: DEBUG nova.network.neutron [-] [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 633.132259] env[63088]: INFO nova.compute.manager [None req-79832ead-48a0-4aad-a47e-8e38c2a687da tempest-ServersTestJSON-106593364 tempest-ServersTestJSON-106593364-project-member] [instance: 34c01c2c-519b-44a1-aba2-c26e2c0b4737] Took 1.18 seconds to deallocate network for instance. [ 633.399923] env[63088]: DEBUG nova.network.neutron [None req-a9773ab2-f8b9-4102-8e41-5ee881a8637e tempest-FloatingIPsAssociationNegativeTestJSON-189613961 tempest-FloatingIPsAssociationNegativeTestJSON-189613961-project-member] [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 633.511557] env[63088]: DEBUG nova.network.neutron [None req-a9773ab2-f8b9-4102-8e41-5ee881a8637e tempest-FloatingIPsAssociationNegativeTestJSON-189613961 tempest-FloatingIPsAssociationNegativeTestJSON-189613961-project-member] [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 633.595583] env[63088]: DEBUG nova.network.neutron [-] [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 633.815612] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c993209-5154-45e7-b7d8-c4189e4b4e66 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.824780] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e3a8a03-8845-4475-9377-a196dff3c53c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.862766] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-319ed9fc-5168-469a-b112-944821cf67da {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.873423] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb8ae60a-68c0-486c-9e6d-ae714319775e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.888678] env[63088]: DEBUG nova.compute.provider_tree [None req-bd71c985-753f-4eb9-b297-7d458ea844aa tempest-ImagesOneServerNegativeTestJSON-1975365234 tempest-ImagesOneServerNegativeTestJSON-1975365234-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 634.017655] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a9773ab2-f8b9-4102-8e41-5ee881a8637e tempest-FloatingIPsAssociationNegativeTestJSON-189613961 tempest-FloatingIPsAssociationNegativeTestJSON-189613961-project-member] Releasing lock "refresh_cache-9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 634.017896] env[63088]: DEBUG nova.compute.manager [None req-a9773ab2-f8b9-4102-8e41-5ee881a8637e tempest-FloatingIPsAssociationNegativeTestJSON-189613961 tempest-FloatingIPsAssociationNegativeTestJSON-189613961-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 634.018125] env[63088]: DEBUG nova.compute.manager [None req-a9773ab2-f8b9-4102-8e41-5ee881a8637e tempest-FloatingIPsAssociationNegativeTestJSON-189613961 tempest-FloatingIPsAssociationNegativeTestJSON-189613961-project-member] [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 634.018371] env[63088]: DEBUG nova.network.neutron [None req-a9773ab2-f8b9-4102-8e41-5ee881a8637e tempest-FloatingIPsAssociationNegativeTestJSON-189613961 tempest-FloatingIPsAssociationNegativeTestJSON-189613961-project-member] [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 634.042068] env[63088]: DEBUG nova.network.neutron [None req-a9773ab2-f8b9-4102-8e41-5ee881a8637e tempest-FloatingIPsAssociationNegativeTestJSON-189613961 tempest-FloatingIPsAssociationNegativeTestJSON-189613961-project-member] [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 634.103505] env[63088]: INFO nova.compute.manager [-] [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] Took 1.03 seconds to deallocate network for instance. [ 634.107098] env[63088]: DEBUG nova.compute.claims [None req-f81b332d-72eb-4149-a6a2-3c8a9c9bec53 tempest-ServersNegativeTestJSON-1008333314 tempest-ServersNegativeTestJSON-1008333314-project-member] [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] Aborting claim: {{(pid=63088) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 634.107521] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f81b332d-72eb-4149-a6a2-3c8a9c9bec53 tempest-ServersNegativeTestJSON-1008333314 tempest-ServersNegativeTestJSON-1008333314-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 634.176409] env[63088]: INFO nova.scheduler.client.report [None req-79832ead-48a0-4aad-a47e-8e38c2a687da tempest-ServersTestJSON-106593364 tempest-ServersTestJSON-106593364-project-member] Deleted allocations for instance 34c01c2c-519b-44a1-aba2-c26e2c0b4737 [ 634.396595] env[63088]: DEBUG nova.scheduler.client.report [None req-bd71c985-753f-4eb9-b297-7d458ea844aa tempest-ImagesOneServerNegativeTestJSON-1975365234 tempest-ImagesOneServerNegativeTestJSON-1975365234-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 634.545988] env[63088]: DEBUG nova.network.neutron [None req-a9773ab2-f8b9-4102-8e41-5ee881a8637e tempest-FloatingIPsAssociationNegativeTestJSON-189613961 tempest-FloatingIPsAssociationNegativeTestJSON-189613961-project-member] [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 634.686121] env[63088]: DEBUG oslo_concurrency.lockutils [None req-79832ead-48a0-4aad-a47e-8e38c2a687da tempest-ServersTestJSON-106593364 tempest-ServersTestJSON-106593364-project-member] Lock "34c01c2c-519b-44a1-aba2-c26e2c0b4737" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 71.792s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 634.901307] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bd71c985-753f-4eb9-b297-7d458ea844aa tempest-ImagesOneServerNegativeTestJSON-1975365234 tempest-ImagesOneServerNegativeTestJSON-1975365234-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.030s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 634.901920] env[63088]: ERROR nova.compute.manager [None req-bd71c985-753f-4eb9-b297-7d458ea844aa tempest-ImagesOneServerNegativeTestJSON-1975365234 tempest-ImagesOneServerNegativeTestJSON-1975365234-project-member] [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c84089e5-c386-4714-aa1d-ef6063e26260, please check neutron logs for more information. [ 634.901920] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] Traceback (most recent call last): [ 634.901920] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 634.901920] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] self.driver.spawn(context, instance, image_meta, [ 634.901920] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 634.901920] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 634.901920] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 634.901920] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] vm_ref = self.build_virtual_machine(instance, [ 634.901920] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 634.901920] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] vif_infos = vmwarevif.get_vif_info(self._session, [ 634.901920] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 634.902203] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] for vif in network_info: [ 634.902203] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 634.902203] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] return self._sync_wrapper(fn, *args, **kwargs) [ 634.902203] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 634.902203] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] self.wait() [ 634.902203] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 634.902203] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] self[:] = self._gt.wait() [ 634.902203] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 634.902203] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] return self._exit_event.wait() [ 634.902203] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 634.902203] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] current.throw(*self._exc) [ 634.902203] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 634.902203] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] result = function(*args, **kwargs) [ 634.902504] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 634.902504] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] return func(*args, **kwargs) [ 634.902504] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 634.902504] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] raise e [ 634.902504] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 634.902504] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] nwinfo = self.network_api.allocate_for_instance( [ 634.902504] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 634.902504] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] created_port_ids = self._update_ports_for_instance( [ 634.902504] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 634.902504] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] with excutils.save_and_reraise_exception(): [ 634.902504] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 634.902504] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] self.force_reraise() [ 634.902504] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 634.902850] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] raise self.value [ 634.902850] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 634.902850] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] updated_port = self._update_port( [ 634.902850] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 634.902850] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] _ensure_no_port_binding_failure(port) [ 634.902850] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 634.902850] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] raise exception.PortBindingFailed(port_id=port['id']) [ 634.902850] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] nova.exception.PortBindingFailed: Binding failed for port c84089e5-c386-4714-aa1d-ef6063e26260, please check neutron logs for more information. [ 634.902850] env[63088]: ERROR nova.compute.manager [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] [ 634.902850] env[63088]: DEBUG nova.compute.utils [None req-bd71c985-753f-4eb9-b297-7d458ea844aa tempest-ImagesOneServerNegativeTestJSON-1975365234 tempest-ImagesOneServerNegativeTestJSON-1975365234-project-member] [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] Binding failed for port c84089e5-c386-4714-aa1d-ef6063e26260, please check neutron logs for more information. {{(pid=63088) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 634.903855] env[63088]: DEBUG oslo_concurrency.lockutils [None req-83cac04a-fc75-40e5-8cdf-832028033c8d tempest-ServersWithSpecificFlavorTestJSON-501634894 tempest-ServersWithSpecificFlavorTestJSON-501634894-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.713s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 634.906957] env[63088]: DEBUG nova.compute.manager [None req-bd71c985-753f-4eb9-b297-7d458ea844aa tempest-ImagesOneServerNegativeTestJSON-1975365234 tempest-ImagesOneServerNegativeTestJSON-1975365234-project-member] [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] Build of instance ca92ddb4-cac8-450a-9b7e-d09ad082354a was re-scheduled: Binding failed for port c84089e5-c386-4714-aa1d-ef6063e26260, please check neutron logs for more information. {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 634.907378] env[63088]: DEBUG nova.compute.manager [None req-bd71c985-753f-4eb9-b297-7d458ea844aa tempest-ImagesOneServerNegativeTestJSON-1975365234 tempest-ImagesOneServerNegativeTestJSON-1975365234-project-member] [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] Unplugging VIFs for instance {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 634.907602] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bd71c985-753f-4eb9-b297-7d458ea844aa tempest-ImagesOneServerNegativeTestJSON-1975365234 tempest-ImagesOneServerNegativeTestJSON-1975365234-project-member] Acquiring lock "refresh_cache-ca92ddb4-cac8-450a-9b7e-d09ad082354a" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 634.907745] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bd71c985-753f-4eb9-b297-7d458ea844aa tempest-ImagesOneServerNegativeTestJSON-1975365234 tempest-ImagesOneServerNegativeTestJSON-1975365234-project-member] Acquired lock "refresh_cache-ca92ddb4-cac8-450a-9b7e-d09ad082354a" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 634.907902] env[63088]: DEBUG nova.network.neutron [None req-bd71c985-753f-4eb9-b297-7d458ea844aa tempest-ImagesOneServerNegativeTestJSON-1975365234 tempest-ImagesOneServerNegativeTestJSON-1975365234-project-member] [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 635.047894] env[63088]: INFO nova.compute.manager [None req-a9773ab2-f8b9-4102-8e41-5ee881a8637e tempest-FloatingIPsAssociationNegativeTestJSON-189613961 tempest-FloatingIPsAssociationNegativeTestJSON-189613961-project-member] [instance: 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4] Took 1.03 seconds to deallocate network for instance. [ 635.143792] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d012c7de-73d3-432e-8cc9-9d2f0456777d tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Acquiring lock "641eed47-54f6-4c08-a8fd-f06bc5f7fb56" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 635.143792] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d012c7de-73d3-432e-8cc9-9d2f0456777d tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Lock "641eed47-54f6-4c08-a8fd-f06bc5f7fb56" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 635.189413] env[63088]: DEBUG nova.compute.manager [None req-224c474a-3c21-4a3f-9526-5bb4e3ba7d78 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 635.437259] env[63088]: DEBUG nova.network.neutron [None req-bd71c985-753f-4eb9-b297-7d458ea844aa tempest-ImagesOneServerNegativeTestJSON-1975365234 tempest-ImagesOneServerNegativeTestJSON-1975365234-project-member] [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 635.530759] env[63088]: DEBUG nova.network.neutron [None req-bd71c985-753f-4eb9-b297-7d458ea844aa tempest-ImagesOneServerNegativeTestJSON-1975365234 tempest-ImagesOneServerNegativeTestJSON-1975365234-project-member] [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 635.708660] env[63088]: DEBUG oslo_concurrency.lockutils [None req-224c474a-3c21-4a3f-9526-5bb4e3ba7d78 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 635.807515] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0864e6d-1acb-4ac3-a3f8-f24464c6c61c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.816234] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e5acb58-ae42-4679-b26d-cbc8b19a35af {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.848962] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-308a3a49-ff46-46ec-b3e4-4fe8e82ebf2b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.857707] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1c256b0-1b7f-4138-a481-d1ef8a784e64 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.873388] env[63088]: DEBUG nova.compute.provider_tree [None req-83cac04a-fc75-40e5-8cdf-832028033c8d tempest-ServersWithSpecificFlavorTestJSON-501634894 tempest-ServersWithSpecificFlavorTestJSON-501634894-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 636.034280] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bd71c985-753f-4eb9-b297-7d458ea844aa tempest-ImagesOneServerNegativeTestJSON-1975365234 tempest-ImagesOneServerNegativeTestJSON-1975365234-project-member] Releasing lock "refresh_cache-ca92ddb4-cac8-450a-9b7e-d09ad082354a" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 636.034545] env[63088]: DEBUG nova.compute.manager [None req-bd71c985-753f-4eb9-b297-7d458ea844aa tempest-ImagesOneServerNegativeTestJSON-1975365234 tempest-ImagesOneServerNegativeTestJSON-1975365234-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 636.034994] env[63088]: DEBUG nova.compute.manager [None req-bd71c985-753f-4eb9-b297-7d458ea844aa tempest-ImagesOneServerNegativeTestJSON-1975365234 tempest-ImagesOneServerNegativeTestJSON-1975365234-project-member] [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 636.034994] env[63088]: DEBUG nova.network.neutron [None req-bd71c985-753f-4eb9-b297-7d458ea844aa tempest-ImagesOneServerNegativeTestJSON-1975365234 tempest-ImagesOneServerNegativeTestJSON-1975365234-project-member] [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 636.054246] env[63088]: DEBUG nova.network.neutron [None req-bd71c985-753f-4eb9-b297-7d458ea844aa tempest-ImagesOneServerNegativeTestJSON-1975365234 tempest-ImagesOneServerNegativeTestJSON-1975365234-project-member] [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 636.078680] env[63088]: INFO nova.scheduler.client.report [None req-a9773ab2-f8b9-4102-8e41-5ee881a8637e tempest-FloatingIPsAssociationNegativeTestJSON-189613961 tempest-FloatingIPsAssociationNegativeTestJSON-189613961-project-member] Deleted allocations for instance 9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4 [ 636.377961] env[63088]: DEBUG nova.scheduler.client.report [None req-83cac04a-fc75-40e5-8cdf-832028033c8d tempest-ServersWithSpecificFlavorTestJSON-501634894 tempest-ServersWithSpecificFlavorTestJSON-501634894-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 636.534099] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bce00a2e-7fb0-44a9-a66e-261b7c3d5384 tempest-ServerActionsV293TestJSON-1088890325 tempest-ServerActionsV293TestJSON-1088890325-project-member] Acquiring lock "b82c89ad-f7d6-4c04-a48e-acdab29ea5e7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 636.534330] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bce00a2e-7fb0-44a9-a66e-261b7c3d5384 tempest-ServerActionsV293TestJSON-1088890325 tempest-ServerActionsV293TestJSON-1088890325-project-member] Lock "b82c89ad-f7d6-4c04-a48e-acdab29ea5e7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 636.559164] env[63088]: DEBUG nova.network.neutron [None req-bd71c985-753f-4eb9-b297-7d458ea844aa tempest-ImagesOneServerNegativeTestJSON-1975365234 tempest-ImagesOneServerNegativeTestJSON-1975365234-project-member] [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 636.586644] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a9773ab2-f8b9-4102-8e41-5ee881a8637e tempest-FloatingIPsAssociationNegativeTestJSON-189613961 tempest-FloatingIPsAssociationNegativeTestJSON-189613961-project-member] Lock "9b7c2500-b6c1-4a82-bbbf-2a23139ddcb4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 72.124s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 636.883649] env[63088]: DEBUG oslo_concurrency.lockutils [None req-83cac04a-fc75-40e5-8cdf-832028033c8d tempest-ServersWithSpecificFlavorTestJSON-501634894 tempest-ServersWithSpecificFlavorTestJSON-501634894-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.980s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 636.884512] env[63088]: ERROR nova.compute.manager [None req-83cac04a-fc75-40e5-8cdf-832028033c8d tempest-ServersWithSpecificFlavorTestJSON-501634894 tempest-ServersWithSpecificFlavorTestJSON-501634894-project-member] [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 228daa45-7c0d-4ae7-a3b3-1887b319e67d, please check neutron logs for more information. [ 636.884512] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] Traceback (most recent call last): [ 636.884512] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 636.884512] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] self.driver.spawn(context, instance, image_meta, [ 636.884512] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 636.884512] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 636.884512] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 636.884512] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] vm_ref = self.build_virtual_machine(instance, [ 636.884512] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 636.884512] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] vif_infos = vmwarevif.get_vif_info(self._session, [ 636.884512] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 636.884974] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] for vif in network_info: [ 636.884974] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 636.884974] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] return self._sync_wrapper(fn, *args, **kwargs) [ 636.884974] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 636.884974] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] self.wait() [ 636.884974] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 636.884974] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] self[:] = self._gt.wait() [ 636.884974] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 636.884974] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] return self._exit_event.wait() [ 636.884974] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 636.884974] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] result = hub.switch() [ 636.884974] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 636.884974] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] return self.greenlet.switch() [ 636.885648] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 636.885648] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] result = function(*args, **kwargs) [ 636.885648] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 636.885648] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] return func(*args, **kwargs) [ 636.885648] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 636.885648] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] raise e [ 636.885648] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 636.885648] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] nwinfo = self.network_api.allocate_for_instance( [ 636.885648] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 636.885648] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] created_port_ids = self._update_ports_for_instance( [ 636.885648] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 636.885648] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] with excutils.save_and_reraise_exception(): [ 636.885648] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 636.886184] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] self.force_reraise() [ 636.886184] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 636.886184] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] raise self.value [ 636.886184] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 636.886184] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] updated_port = self._update_port( [ 636.886184] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 636.886184] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] _ensure_no_port_binding_failure(port) [ 636.886184] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 636.886184] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] raise exception.PortBindingFailed(port_id=port['id']) [ 636.886184] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] nova.exception.PortBindingFailed: Binding failed for port 228daa45-7c0d-4ae7-a3b3-1887b319e67d, please check neutron logs for more information. [ 636.886184] env[63088]: ERROR nova.compute.manager [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] [ 636.886632] env[63088]: DEBUG nova.compute.utils [None req-83cac04a-fc75-40e5-8cdf-832028033c8d tempest-ServersWithSpecificFlavorTestJSON-501634894 tempest-ServersWithSpecificFlavorTestJSON-501634894-project-member] [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] Binding failed for port 228daa45-7c0d-4ae7-a3b3-1887b319e67d, please check neutron logs for more information. {{(pid=63088) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 636.891133] env[63088]: DEBUG oslo_concurrency.lockutils [None req-acfde423-bb82-457b-a17c-3c463abca710 tempest-ServersTestFqdnHostnames-1568147653 tempest-ServersTestFqdnHostnames-1568147653-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 21.386s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 636.895628] env[63088]: DEBUG nova.compute.manager [None req-83cac04a-fc75-40e5-8cdf-832028033c8d tempest-ServersWithSpecificFlavorTestJSON-501634894 tempest-ServersWithSpecificFlavorTestJSON-501634894-project-member] [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] Build of instance d6c76851-7059-40a4-a39b-e625c99dfcd9 was re-scheduled: Binding failed for port 228daa45-7c0d-4ae7-a3b3-1887b319e67d, please check neutron logs for more information. {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 636.896367] env[63088]: DEBUG nova.compute.manager [None req-83cac04a-fc75-40e5-8cdf-832028033c8d tempest-ServersWithSpecificFlavorTestJSON-501634894 tempest-ServersWithSpecificFlavorTestJSON-501634894-project-member] [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] Unplugging VIFs for instance {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 636.897364] env[63088]: DEBUG oslo_concurrency.lockutils [None req-83cac04a-fc75-40e5-8cdf-832028033c8d tempest-ServersWithSpecificFlavorTestJSON-501634894 tempest-ServersWithSpecificFlavorTestJSON-501634894-project-member] Acquiring lock "refresh_cache-d6c76851-7059-40a4-a39b-e625c99dfcd9" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 636.898757] env[63088]: DEBUG oslo_concurrency.lockutils [None req-83cac04a-fc75-40e5-8cdf-832028033c8d tempest-ServersWithSpecificFlavorTestJSON-501634894 tempest-ServersWithSpecificFlavorTestJSON-501634894-project-member] Acquired lock "refresh_cache-d6c76851-7059-40a4-a39b-e625c99dfcd9" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 636.898951] env[63088]: DEBUG nova.network.neutron [None req-83cac04a-fc75-40e5-8cdf-832028033c8d tempest-ServersWithSpecificFlavorTestJSON-501634894 tempest-ServersWithSpecificFlavorTestJSON-501634894-project-member] [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 637.062237] env[63088]: INFO nova.compute.manager [None req-bd71c985-753f-4eb9-b297-7d458ea844aa tempest-ImagesOneServerNegativeTestJSON-1975365234 tempest-ImagesOneServerNegativeTestJSON-1975365234-project-member] [instance: ca92ddb4-cac8-450a-9b7e-d09ad082354a] Took 1.03 seconds to deallocate network for instance. [ 637.088957] env[63088]: DEBUG nova.compute.manager [None req-cab9e1ae-5fd3-4816-8684-8d11c94f36d2 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 637.431905] env[63088]: DEBUG nova.network.neutron [None req-83cac04a-fc75-40e5-8cdf-832028033c8d tempest-ServersWithSpecificFlavorTestJSON-501634894 tempest-ServersWithSpecificFlavorTestJSON-501634894-project-member] [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 637.604207] env[63088]: DEBUG nova.network.neutron [None req-83cac04a-fc75-40e5-8cdf-832028033c8d tempest-ServersWithSpecificFlavorTestJSON-501634894 tempest-ServersWithSpecificFlavorTestJSON-501634894-project-member] [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 637.622499] env[63088]: DEBUG oslo_concurrency.lockutils [None req-cab9e1ae-5fd3-4816-8684-8d11c94f36d2 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 637.943634] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e73d4f85-e322-432e-a5a8-90ad1315b1db {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.952712] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e3db0ea-a426-49e7-886c-bb71a1392ded {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.986027] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e7c2cd2-24dc-4ba4-bfe7-445e9eb4b66c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.994685] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e17430c-5e44-4585-825b-4e8d6c55d5fc {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.010054] env[63088]: DEBUG nova.compute.provider_tree [None req-acfde423-bb82-457b-a17c-3c463abca710 tempest-ServersTestFqdnHostnames-1568147653 tempest-ServersTestFqdnHostnames-1568147653-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 638.106037] env[63088]: INFO nova.scheduler.client.report [None req-bd71c985-753f-4eb9-b297-7d458ea844aa tempest-ImagesOneServerNegativeTestJSON-1975365234 tempest-ImagesOneServerNegativeTestJSON-1975365234-project-member] Deleted allocations for instance ca92ddb4-cac8-450a-9b7e-d09ad082354a [ 638.113405] env[63088]: DEBUG oslo_concurrency.lockutils [None req-83cac04a-fc75-40e5-8cdf-832028033c8d tempest-ServersWithSpecificFlavorTestJSON-501634894 tempest-ServersWithSpecificFlavorTestJSON-501634894-project-member] Releasing lock "refresh_cache-d6c76851-7059-40a4-a39b-e625c99dfcd9" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 638.113597] env[63088]: DEBUG nova.compute.manager [None req-83cac04a-fc75-40e5-8cdf-832028033c8d tempest-ServersWithSpecificFlavorTestJSON-501634894 tempest-ServersWithSpecificFlavorTestJSON-501634894-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 638.113780] env[63088]: DEBUG nova.compute.manager [None req-83cac04a-fc75-40e5-8cdf-832028033c8d tempest-ServersWithSpecificFlavorTestJSON-501634894 tempest-ServersWithSpecificFlavorTestJSON-501634894-project-member] [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 638.114018] env[63088]: DEBUG nova.network.neutron [None req-83cac04a-fc75-40e5-8cdf-832028033c8d tempest-ServersWithSpecificFlavorTestJSON-501634894 tempest-ServersWithSpecificFlavorTestJSON-501634894-project-member] [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 638.135671] env[63088]: DEBUG nova.network.neutron [None req-83cac04a-fc75-40e5-8cdf-832028033c8d tempest-ServersWithSpecificFlavorTestJSON-501634894 tempest-ServersWithSpecificFlavorTestJSON-501634894-project-member] [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 638.515292] env[63088]: DEBUG nova.scheduler.client.report [None req-acfde423-bb82-457b-a17c-3c463abca710 tempest-ServersTestFqdnHostnames-1568147653 tempest-ServersTestFqdnHostnames-1568147653-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 638.617994] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bd71c985-753f-4eb9-b297-7d458ea844aa tempest-ImagesOneServerNegativeTestJSON-1975365234 tempest-ImagesOneServerNegativeTestJSON-1975365234-project-member] Lock "ca92ddb4-cac8-450a-9b7e-d09ad082354a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 72.140s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 638.641101] env[63088]: DEBUG nova.network.neutron [None req-83cac04a-fc75-40e5-8cdf-832028033c8d tempest-ServersWithSpecificFlavorTestJSON-501634894 tempest-ServersWithSpecificFlavorTestJSON-501634894-project-member] [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 639.021674] env[63088]: DEBUG oslo_concurrency.lockutils [None req-acfde423-bb82-457b-a17c-3c463abca710 tempest-ServersTestFqdnHostnames-1568147653 tempest-ServersTestFqdnHostnames-1568147653-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.130s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 639.022800] env[63088]: ERROR nova.compute.manager [None req-acfde423-bb82-457b-a17c-3c463abca710 tempest-ServersTestFqdnHostnames-1568147653 tempest-ServersTestFqdnHostnames-1568147653-project-member] [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c5aeb7be-57db-42c6-81aa-bbcdcef2ce85, please check neutron logs for more information. [ 639.022800] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] Traceback (most recent call last): [ 639.022800] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 639.022800] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] self.driver.spawn(context, instance, image_meta, [ 639.022800] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 639.022800] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] self._vmops.spawn(context, instance, image_meta, injected_files, [ 639.022800] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 639.022800] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] vm_ref = self.build_virtual_machine(instance, [ 639.022800] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 639.022800] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] vif_infos = vmwarevif.get_vif_info(self._session, [ 639.022800] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 639.023212] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] for vif in network_info: [ 639.023212] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 639.023212] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] return self._sync_wrapper(fn, *args, **kwargs) [ 639.023212] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 639.023212] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] self.wait() [ 639.023212] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 639.023212] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] self[:] = self._gt.wait() [ 639.023212] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 639.023212] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] return self._exit_event.wait() [ 639.023212] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 639.023212] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] current.throw(*self._exc) [ 639.023212] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 639.023212] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] result = function(*args, **kwargs) [ 639.023557] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 639.023557] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] return func(*args, **kwargs) [ 639.023557] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 639.023557] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] raise e [ 639.023557] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 639.023557] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] nwinfo = self.network_api.allocate_for_instance( [ 639.023557] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 639.023557] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] created_port_ids = self._update_ports_for_instance( [ 639.023557] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 639.023557] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] with excutils.save_and_reraise_exception(): [ 639.023557] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 639.023557] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] self.force_reraise() [ 639.023557] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 639.023861] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] raise self.value [ 639.023861] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 639.023861] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] updated_port = self._update_port( [ 639.023861] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 639.023861] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] _ensure_no_port_binding_failure(port) [ 639.023861] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 639.023861] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] raise exception.PortBindingFailed(port_id=port['id']) [ 639.023861] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] nova.exception.PortBindingFailed: Binding failed for port c5aeb7be-57db-42c6-81aa-bbcdcef2ce85, please check neutron logs for more information. [ 639.023861] env[63088]: ERROR nova.compute.manager [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] [ 639.023861] env[63088]: DEBUG nova.compute.utils [None req-acfde423-bb82-457b-a17c-3c463abca710 tempest-ServersTestFqdnHostnames-1568147653 tempest-ServersTestFqdnHostnames-1568147653-project-member] [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] Binding failed for port c5aeb7be-57db-42c6-81aa-bbcdcef2ce85, please check neutron logs for more information. {{(pid=63088) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 639.026215] env[63088]: DEBUG nova.compute.manager [None req-acfde423-bb82-457b-a17c-3c463abca710 tempest-ServersTestFqdnHostnames-1568147653 tempest-ServersTestFqdnHostnames-1568147653-project-member] [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] Build of instance 91a09343-c2d7-4f5d-ae0a-4a123af92fba was re-scheduled: Binding failed for port c5aeb7be-57db-42c6-81aa-bbcdcef2ce85, please check neutron logs for more information. {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 639.026692] env[63088]: DEBUG nova.compute.manager [None req-acfde423-bb82-457b-a17c-3c463abca710 tempest-ServersTestFqdnHostnames-1568147653 tempest-ServersTestFqdnHostnames-1568147653-project-member] [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] Unplugging VIFs for instance {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 639.026924] env[63088]: DEBUG oslo_concurrency.lockutils [None req-acfde423-bb82-457b-a17c-3c463abca710 tempest-ServersTestFqdnHostnames-1568147653 tempest-ServersTestFqdnHostnames-1568147653-project-member] Acquiring lock "refresh_cache-91a09343-c2d7-4f5d-ae0a-4a123af92fba" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 639.027073] env[63088]: DEBUG oslo_concurrency.lockutils [None req-acfde423-bb82-457b-a17c-3c463abca710 tempest-ServersTestFqdnHostnames-1568147653 tempest-ServersTestFqdnHostnames-1568147653-project-member] Acquired lock "refresh_cache-91a09343-c2d7-4f5d-ae0a-4a123af92fba" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 639.027234] env[63088]: DEBUG nova.network.neutron [None req-acfde423-bb82-457b-a17c-3c463abca710 tempest-ServersTestFqdnHostnames-1568147653 tempest-ServersTestFqdnHostnames-1568147653-project-member] [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 639.028688] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f5f50761-0430-407c-84cc-7b2c9cf8683b tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.907s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 639.120374] env[63088]: DEBUG nova.compute.manager [None req-b7ea81d3-fe92-4160-8c6b-de9afb157d35 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 639.144944] env[63088]: INFO nova.compute.manager [None req-83cac04a-fc75-40e5-8cdf-832028033c8d tempest-ServersWithSpecificFlavorTestJSON-501634894 tempest-ServersWithSpecificFlavorTestJSON-501634894-project-member] [instance: d6c76851-7059-40a4-a39b-e625c99dfcd9] Took 1.03 seconds to deallocate network for instance. [ 639.555126] env[63088]: DEBUG nova.network.neutron [None req-acfde423-bb82-457b-a17c-3c463abca710 tempest-ServersTestFqdnHostnames-1568147653 tempest-ServersTestFqdnHostnames-1568147653-project-member] [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 639.646068] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b7ea81d3-fe92-4160-8c6b-de9afb157d35 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 639.919978] env[63088]: DEBUG nova.network.neutron [None req-acfde423-bb82-457b-a17c-3c463abca710 tempest-ServersTestFqdnHostnames-1568147653 tempest-ServersTestFqdnHostnames-1568147653-project-member] [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 639.953809] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2193b3e-bb1d-4965-ac75-620e637d0cd7 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.962466] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b1fb9f9-ca52-4cd5-a47c-0c3778e0d298 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.004616] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93491e0c-7a9f-470d-be80-c309eea5e723 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.014276] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24aa6ebd-44d2-4486-9342-11b675949280 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.029950] env[63088]: DEBUG nova.compute.provider_tree [None req-f5f50761-0430-407c-84cc-7b2c9cf8683b tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 640.199066] env[63088]: INFO nova.scheduler.client.report [None req-83cac04a-fc75-40e5-8cdf-832028033c8d tempest-ServersWithSpecificFlavorTestJSON-501634894 tempest-ServersWithSpecificFlavorTestJSON-501634894-project-member] Deleted allocations for instance d6c76851-7059-40a4-a39b-e625c99dfcd9 [ 640.424479] env[63088]: DEBUG oslo_concurrency.lockutils [None req-acfde423-bb82-457b-a17c-3c463abca710 tempest-ServersTestFqdnHostnames-1568147653 tempest-ServersTestFqdnHostnames-1568147653-project-member] Releasing lock "refresh_cache-91a09343-c2d7-4f5d-ae0a-4a123af92fba" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 640.424732] env[63088]: DEBUG nova.compute.manager [None req-acfde423-bb82-457b-a17c-3c463abca710 tempest-ServersTestFqdnHostnames-1568147653 tempest-ServersTestFqdnHostnames-1568147653-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 640.424922] env[63088]: DEBUG nova.compute.manager [None req-acfde423-bb82-457b-a17c-3c463abca710 tempest-ServersTestFqdnHostnames-1568147653 tempest-ServersTestFqdnHostnames-1568147653-project-member] [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 640.425103] env[63088]: DEBUG nova.network.neutron [None req-acfde423-bb82-457b-a17c-3c463abca710 tempest-ServersTestFqdnHostnames-1568147653 tempest-ServersTestFqdnHostnames-1568147653-project-member] [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 640.468279] env[63088]: DEBUG nova.network.neutron [None req-acfde423-bb82-457b-a17c-3c463abca710 tempest-ServersTestFqdnHostnames-1568147653 tempest-ServersTestFqdnHostnames-1568147653-project-member] [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 640.532884] env[63088]: DEBUG nova.scheduler.client.report [None req-f5f50761-0430-407c-84cc-7b2c9cf8683b tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 640.708570] env[63088]: DEBUG oslo_concurrency.lockutils [None req-83cac04a-fc75-40e5-8cdf-832028033c8d tempest-ServersWithSpecificFlavorTestJSON-501634894 tempest-ServersWithSpecificFlavorTestJSON-501634894-project-member] Lock "d6c76851-7059-40a4-a39b-e625c99dfcd9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 71.911s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 640.971263] env[63088]: DEBUG nova.network.neutron [None req-acfde423-bb82-457b-a17c-3c463abca710 tempest-ServersTestFqdnHostnames-1568147653 tempest-ServersTestFqdnHostnames-1568147653-project-member] [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 641.038654] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f5f50761-0430-407c-84cc-7b2c9cf8683b tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.010s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 641.039317] env[63088]: ERROR nova.compute.manager [None req-f5f50761-0430-407c-84cc-7b2c9cf8683b tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f6eeb038-e18c-4985-82b2-4297f9655e25, please check neutron logs for more information. [ 641.039317] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] Traceback (most recent call last): [ 641.039317] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 641.039317] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] self.driver.spawn(context, instance, image_meta, [ 641.039317] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 641.039317] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] self._vmops.spawn(context, instance, image_meta, injected_files, [ 641.039317] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 641.039317] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] vm_ref = self.build_virtual_machine(instance, [ 641.039317] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 641.039317] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] vif_infos = vmwarevif.get_vif_info(self._session, [ 641.039317] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 641.039663] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] for vif in network_info: [ 641.039663] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 641.039663] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] return self._sync_wrapper(fn, *args, **kwargs) [ 641.039663] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 641.039663] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] self.wait() [ 641.039663] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 641.039663] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] self[:] = self._gt.wait() [ 641.039663] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 641.039663] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] return self._exit_event.wait() [ 641.039663] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 641.039663] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] result = hub.switch() [ 641.039663] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 641.039663] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] return self.greenlet.switch() [ 641.040042] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 641.040042] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] result = function(*args, **kwargs) [ 641.040042] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 641.040042] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] return func(*args, **kwargs) [ 641.040042] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 641.040042] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] raise e [ 641.040042] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 641.040042] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] nwinfo = self.network_api.allocate_for_instance( [ 641.040042] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 641.040042] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] created_port_ids = self._update_ports_for_instance( [ 641.040042] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 641.040042] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] with excutils.save_and_reraise_exception(): [ 641.040042] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 641.040885] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] self.force_reraise() [ 641.040885] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 641.040885] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] raise self.value [ 641.040885] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 641.040885] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] updated_port = self._update_port( [ 641.040885] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 641.040885] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] _ensure_no_port_binding_failure(port) [ 641.040885] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 641.040885] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] raise exception.PortBindingFailed(port_id=port['id']) [ 641.040885] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] nova.exception.PortBindingFailed: Binding failed for port f6eeb038-e18c-4985-82b2-4297f9655e25, please check neutron logs for more information. [ 641.040885] env[63088]: ERROR nova.compute.manager [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] [ 641.041390] env[63088]: DEBUG nova.compute.utils [None req-f5f50761-0430-407c-84cc-7b2c9cf8683b tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] Binding failed for port f6eeb038-e18c-4985-82b2-4297f9655e25, please check neutron logs for more information. {{(pid=63088) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 641.042402] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a5e283fa-bcc1-4186-a4e3-1ae9f8d221af tempest-ServerMetadataTestJSON-859565163 tempest-ServerMetadataTestJSON-859565163-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.376s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 641.043871] env[63088]: INFO nova.compute.claims [None req-a5e283fa-bcc1-4186-a4e3-1ae9f8d221af tempest-ServerMetadataTestJSON-859565163 tempest-ServerMetadataTestJSON-859565163-project-member] [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 641.046824] env[63088]: DEBUG nova.compute.manager [None req-f5f50761-0430-407c-84cc-7b2c9cf8683b tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] Build of instance f29f6a5e-8621-4802-a679-d5c9a8fb2461 was re-scheduled: Binding failed for port f6eeb038-e18c-4985-82b2-4297f9655e25, please check neutron logs for more information. {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 641.047464] env[63088]: DEBUG nova.compute.manager [None req-f5f50761-0430-407c-84cc-7b2c9cf8683b tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] Unplugging VIFs for instance {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 641.047700] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f5f50761-0430-407c-84cc-7b2c9cf8683b tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] Acquiring lock "refresh_cache-f29f6a5e-8621-4802-a679-d5c9a8fb2461" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 641.047991] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f5f50761-0430-407c-84cc-7b2c9cf8683b tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] Acquired lock "refresh_cache-f29f6a5e-8621-4802-a679-d5c9a8fb2461" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 641.047991] env[63088]: DEBUG nova.network.neutron [None req-f5f50761-0430-407c-84cc-7b2c9cf8683b tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 641.211244] env[63088]: DEBUG nova.compute.manager [None req-442e0a41-8be2-46c5-9922-69b38e5ecc58 tempest-ServerAddressesNegativeTestJSON-62458283 tempest-ServerAddressesNegativeTestJSON-62458283-project-member] [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 641.474739] env[63088]: INFO nova.compute.manager [None req-acfde423-bb82-457b-a17c-3c463abca710 tempest-ServersTestFqdnHostnames-1568147653 tempest-ServersTestFqdnHostnames-1568147653-project-member] [instance: 91a09343-c2d7-4f5d-ae0a-4a123af92fba] Took 1.05 seconds to deallocate network for instance. [ 641.578631] env[63088]: DEBUG nova.network.neutron [None req-f5f50761-0430-407c-84cc-7b2c9cf8683b tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 641.740349] env[63088]: DEBUG oslo_concurrency.lockutils [None req-442e0a41-8be2-46c5-9922-69b38e5ecc58 tempest-ServerAddressesNegativeTestJSON-62458283 tempest-ServerAddressesNegativeTestJSON-62458283-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 641.950830] env[63088]: DEBUG nova.network.neutron [None req-f5f50761-0430-407c-84cc-7b2c9cf8683b tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 642.453171] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f5f50761-0430-407c-84cc-7b2c9cf8683b tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] Releasing lock "refresh_cache-f29f6a5e-8621-4802-a679-d5c9a8fb2461" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 642.453455] env[63088]: DEBUG nova.compute.manager [None req-f5f50761-0430-407c-84cc-7b2c9cf8683b tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 642.453578] env[63088]: DEBUG nova.compute.manager [None req-f5f50761-0430-407c-84cc-7b2c9cf8683b tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 642.453736] env[63088]: DEBUG nova.network.neutron [None req-f5f50761-0430-407c-84cc-7b2c9cf8683b tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 642.479421] env[63088]: DEBUG nova.network.neutron [None req-f5f50761-0430-407c-84cc-7b2c9cf8683b tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 642.514346] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20ed8cbe-752d-4555-9f0d-fb836fcd936e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.522536] env[63088]: INFO nova.scheduler.client.report [None req-acfde423-bb82-457b-a17c-3c463abca710 tempest-ServersTestFqdnHostnames-1568147653 tempest-ServersTestFqdnHostnames-1568147653-project-member] Deleted allocations for instance 91a09343-c2d7-4f5d-ae0a-4a123af92fba [ 642.542234] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d81eb6de-4153-4a75-9d7c-66240324eee3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.579893] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5cc479b-6f57-4180-b6a8-10ecd94bb7d6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.589211] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd0c38e0-f7d2-4837-9c61-8837f538497c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.606500] env[63088]: DEBUG nova.compute.provider_tree [None req-a5e283fa-bcc1-4186-a4e3-1ae9f8d221af tempest-ServerMetadataTestJSON-859565163 tempest-ServerMetadataTestJSON-859565163-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 642.981649] env[63088]: DEBUG nova.network.neutron [None req-f5f50761-0430-407c-84cc-7b2c9cf8683b tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 643.045570] env[63088]: DEBUG oslo_concurrency.lockutils [None req-acfde423-bb82-457b-a17c-3c463abca710 tempest-ServersTestFqdnHostnames-1568147653 tempest-ServersTestFqdnHostnames-1568147653-project-member] Lock "91a09343-c2d7-4f5d-ae0a-4a123af92fba" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 72.937s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 643.110270] env[63088]: DEBUG nova.scheduler.client.report [None req-a5e283fa-bcc1-4186-a4e3-1ae9f8d221af tempest-ServerMetadataTestJSON-859565163 tempest-ServerMetadataTestJSON-859565163-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 643.484966] env[63088]: INFO nova.compute.manager [None req-f5f50761-0430-407c-84cc-7b2c9cf8683b tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] [instance: f29f6a5e-8621-4802-a679-d5c9a8fb2461] Took 1.03 seconds to deallocate network for instance. [ 643.548472] env[63088]: DEBUG nova.compute.manager [None req-32cec7a8-18a7-43df-b82f-d466be0dcd4a tempest-AttachInterfacesUnderV243Test-1205880501 tempest-AttachInterfacesUnderV243Test-1205880501-project-member] [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 643.617686] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a5e283fa-bcc1-4186-a4e3-1ae9f8d221af tempest-ServerMetadataTestJSON-859565163 tempest-ServerMetadataTestJSON-859565163-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.575s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 643.618290] env[63088]: DEBUG nova.compute.manager [None req-a5e283fa-bcc1-4186-a4e3-1ae9f8d221af tempest-ServerMetadataTestJSON-859565163 tempest-ServerMetadataTestJSON-859565163-project-member] [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 643.621868] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b1b885bf-cef9-4808-8f16-707ab04499b3 tempest-ServersTestManualDisk-1859582357 tempest-ServersTestManualDisk-1859582357-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.922s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 644.078509] env[63088]: DEBUG oslo_concurrency.lockutils [None req-32cec7a8-18a7-43df-b82f-d466be0dcd4a tempest-AttachInterfacesUnderV243Test-1205880501 tempest-AttachInterfacesUnderV243Test-1205880501-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 644.125107] env[63088]: DEBUG nova.compute.utils [None req-a5e283fa-bcc1-4186-a4e3-1ae9f8d221af tempest-ServerMetadataTestJSON-859565163 tempest-ServerMetadataTestJSON-859565163-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 644.127741] env[63088]: DEBUG nova.compute.manager [None req-a5e283fa-bcc1-4186-a4e3-1ae9f8d221af tempest-ServerMetadataTestJSON-859565163 tempest-ServerMetadataTestJSON-859565163-project-member] [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 644.128273] env[63088]: DEBUG nova.network.neutron [None req-a5e283fa-bcc1-4186-a4e3-1ae9f8d221af tempest-ServerMetadataTestJSON-859565163 tempest-ServerMetadataTestJSON-859565163-project-member] [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 644.239245] env[63088]: DEBUG nova.policy [None req-a5e283fa-bcc1-4186-a4e3-1ae9f8d221af tempest-ServerMetadataTestJSON-859565163 tempest-ServerMetadataTestJSON-859565163-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f2b23ef8e6db42f88743849ab21478ef', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e7e04b5845594d69958dec5c66a722d3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 644.533832] env[63088]: INFO nova.scheduler.client.report [None req-f5f50761-0430-407c-84cc-7b2c9cf8683b tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] Deleted allocations for instance f29f6a5e-8621-4802-a679-d5c9a8fb2461 [ 645.185039] env[63088]: DEBUG nova.compute.manager [None req-a5e283fa-bcc1-4186-a4e3-1ae9f8d221af tempest-ServerMetadataTestJSON-859565163 tempest-ServerMetadataTestJSON-859565163-project-member] [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 645.190131] env[63088]: DEBUG nova.network.neutron [None req-a5e283fa-bcc1-4186-a4e3-1ae9f8d221af tempest-ServerMetadataTestJSON-859565163 tempest-ServerMetadataTestJSON-859565163-project-member] [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] Successfully created port: 3aed79cc-3d41-4762-992b-05bb123fc735 {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 645.196740] env[63088]: DEBUG oslo_concurrency.lockutils [None req-1874938e-086f-4c3c-9049-55ea0246719c tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] Acquiring lock "1e1732ef-a198-46e7-8f5e-3b9d8fe8c800" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 645.196740] env[63088]: DEBUG oslo_concurrency.lockutils [None req-1874938e-086f-4c3c-9049-55ea0246719c tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] Lock "1e1732ef-a198-46e7-8f5e-3b9d8fe8c800" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 645.266188] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e10b92c-4676-4d13-9dc2-db38bbf5ed41 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.275682] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ac15500-3505-4612-905f-d98dd0cb6bda {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.307754] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad5e8495-bba6-4fd7-adac-0a4369f77daf {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.319019] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa5e8d88-6358-4a2f-9804-0ab04e8f7ce5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.330901] env[63088]: DEBUG nova.compute.provider_tree [None req-b1b885bf-cef9-4808-8f16-707ab04499b3 tempest-ServersTestManualDisk-1859582357 tempest-ServersTestManualDisk-1859582357-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 645.696510] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f5f50761-0430-407c-84cc-7b2c9cf8683b tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] Lock "f29f6a5e-8621-4802-a679-d5c9a8fb2461" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 71.612s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 645.835068] env[63088]: DEBUG nova.scheduler.client.report [None req-b1b885bf-cef9-4808-8f16-707ab04499b3 tempest-ServersTestManualDisk-1859582357 tempest-ServersTestManualDisk-1859582357-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 646.029269] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3edaa40f-e8ca-4f6f-a9f2-5744f490f65d tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] Acquiring lock "17725a7b-ccd3-477e-a7a6-b96af45d6d1f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 646.030569] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3edaa40f-e8ca-4f6f-a9f2-5744f490f65d tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] Lock "17725a7b-ccd3-477e-a7a6-b96af45d6d1f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 646.202101] env[63088]: DEBUG nova.compute.manager [None req-a5e283fa-bcc1-4186-a4e3-1ae9f8d221af tempest-ServerMetadataTestJSON-859565163 tempest-ServerMetadataTestJSON-859565163-project-member] [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 646.204965] env[63088]: DEBUG nova.compute.manager [None req-bb8843aa-5c19-4289-a41c-17cbfcb47a00 tempest-ServerActionsTestOtherA-1614359098 tempest-ServerActionsTestOtherA-1614359098-project-member] [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 646.245769] env[63088]: DEBUG nova.virt.hardware [None req-a5e283fa-bcc1-4186-a4e3-1ae9f8d221af tempest-ServerMetadataTestJSON-859565163 tempest-ServerMetadataTestJSON-859565163-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 646.246217] env[63088]: DEBUG nova.virt.hardware [None req-a5e283fa-bcc1-4186-a4e3-1ae9f8d221af tempest-ServerMetadataTestJSON-859565163 tempest-ServerMetadataTestJSON-859565163-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 646.246492] env[63088]: DEBUG nova.virt.hardware [None req-a5e283fa-bcc1-4186-a4e3-1ae9f8d221af tempest-ServerMetadataTestJSON-859565163 tempest-ServerMetadataTestJSON-859565163-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 646.246908] env[63088]: DEBUG nova.virt.hardware [None req-a5e283fa-bcc1-4186-a4e3-1ae9f8d221af tempest-ServerMetadataTestJSON-859565163 tempest-ServerMetadataTestJSON-859565163-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 646.247184] env[63088]: DEBUG nova.virt.hardware [None req-a5e283fa-bcc1-4186-a4e3-1ae9f8d221af tempest-ServerMetadataTestJSON-859565163 tempest-ServerMetadataTestJSON-859565163-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 646.247431] env[63088]: DEBUG nova.virt.hardware [None req-a5e283fa-bcc1-4186-a4e3-1ae9f8d221af tempest-ServerMetadataTestJSON-859565163 tempest-ServerMetadataTestJSON-859565163-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 646.247995] env[63088]: DEBUG nova.virt.hardware [None req-a5e283fa-bcc1-4186-a4e3-1ae9f8d221af tempest-ServerMetadataTestJSON-859565163 tempest-ServerMetadataTestJSON-859565163-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 646.248291] env[63088]: DEBUG nova.virt.hardware [None req-a5e283fa-bcc1-4186-a4e3-1ae9f8d221af tempest-ServerMetadataTestJSON-859565163 tempest-ServerMetadataTestJSON-859565163-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 646.248602] env[63088]: DEBUG nova.virt.hardware [None req-a5e283fa-bcc1-4186-a4e3-1ae9f8d221af tempest-ServerMetadataTestJSON-859565163 tempest-ServerMetadataTestJSON-859565163-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 646.248878] env[63088]: DEBUG nova.virt.hardware [None req-a5e283fa-bcc1-4186-a4e3-1ae9f8d221af tempest-ServerMetadataTestJSON-859565163 tempest-ServerMetadataTestJSON-859565163-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 646.249166] env[63088]: DEBUG nova.virt.hardware [None req-a5e283fa-bcc1-4186-a4e3-1ae9f8d221af tempest-ServerMetadataTestJSON-859565163 tempest-ServerMetadataTestJSON-859565163-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 646.250161] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f2243b0-d6b0-414b-96d8-cc8c932d088a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.274943] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7ee8de0-c94e-4391-9dda-55cb748af8c6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.348018] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b1b885bf-cef9-4808-8f16-707ab04499b3 tempest-ServersTestManualDisk-1859582357 tempest-ServersTestManualDisk-1859582357-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.724s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 646.348018] env[63088]: ERROR nova.compute.manager [None req-b1b885bf-cef9-4808-8f16-707ab04499b3 tempest-ServersTestManualDisk-1859582357 tempest-ServersTestManualDisk-1859582357-project-member] [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port fc9be7f9-f0c9-43c8-8828-f85621944087, please check neutron logs for more information. [ 646.348018] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] Traceback (most recent call last): [ 646.348018] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 646.348018] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] self.driver.spawn(context, instance, image_meta, [ 646.348018] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 646.348018] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] self._vmops.spawn(context, instance, image_meta, injected_files, [ 646.348018] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 646.348018] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] vm_ref = self.build_virtual_machine(instance, [ 646.348373] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 646.348373] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] vif_infos = vmwarevif.get_vif_info(self._session, [ 646.348373] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 646.348373] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] for vif in network_info: [ 646.348373] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 646.348373] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] return self._sync_wrapper(fn, *args, **kwargs) [ 646.348373] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 646.348373] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] self.wait() [ 646.348373] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 646.348373] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] self[:] = self._gt.wait() [ 646.348373] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 646.348373] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] return self._exit_event.wait() [ 646.348373] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 646.348713] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] result = hub.switch() [ 646.348713] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 646.348713] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] return self.greenlet.switch() [ 646.348713] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 646.348713] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] result = function(*args, **kwargs) [ 646.348713] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 646.348713] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] return func(*args, **kwargs) [ 646.348713] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 646.348713] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] raise e [ 646.348713] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 646.348713] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] nwinfo = self.network_api.allocate_for_instance( [ 646.348713] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 646.348713] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] created_port_ids = self._update_ports_for_instance( [ 646.348995] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 646.348995] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] with excutils.save_and_reraise_exception(): [ 646.348995] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 646.348995] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] self.force_reraise() [ 646.348995] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 646.348995] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] raise self.value [ 646.348995] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 646.348995] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] updated_port = self._update_port( [ 646.348995] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 646.348995] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] _ensure_no_port_binding_failure(port) [ 646.348995] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 646.348995] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] raise exception.PortBindingFailed(port_id=port['id']) [ 646.349264] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] nova.exception.PortBindingFailed: Binding failed for port fc9be7f9-f0c9-43c8-8828-f85621944087, please check neutron logs for more information. [ 646.349264] env[63088]: ERROR nova.compute.manager [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] [ 646.349264] env[63088]: DEBUG nova.compute.utils [None req-b1b885bf-cef9-4808-8f16-707ab04499b3 tempest-ServersTestManualDisk-1859582357 tempest-ServersTestManualDisk-1859582357-project-member] [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] Binding failed for port fc9be7f9-f0c9-43c8-8828-f85621944087, please check neutron logs for more information. {{(pid=63088) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 646.352347] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0564aade-729d-4cf1-a59d-513384386c54 tempest-AttachInterfacesV270Test-1783457293 tempest-AttachInterfacesV270Test-1783457293-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.844s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 646.355159] env[63088]: DEBUG nova.compute.manager [None req-b1b885bf-cef9-4808-8f16-707ab04499b3 tempest-ServersTestManualDisk-1859582357 tempest-ServersTestManualDisk-1859582357-project-member] [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] Build of instance 498c8e1e-18c6-4e09-a90d-106600bed2cd was re-scheduled: Binding failed for port fc9be7f9-f0c9-43c8-8828-f85621944087, please check neutron logs for more information. {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 646.355693] env[63088]: DEBUG nova.compute.manager [None req-b1b885bf-cef9-4808-8f16-707ab04499b3 tempest-ServersTestManualDisk-1859582357 tempest-ServersTestManualDisk-1859582357-project-member] [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] Unplugging VIFs for instance {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 646.355962] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b1b885bf-cef9-4808-8f16-707ab04499b3 tempest-ServersTestManualDisk-1859582357 tempest-ServersTestManualDisk-1859582357-project-member] Acquiring lock "refresh_cache-498c8e1e-18c6-4e09-a90d-106600bed2cd" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 646.356158] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b1b885bf-cef9-4808-8f16-707ab04499b3 tempest-ServersTestManualDisk-1859582357 tempest-ServersTestManualDisk-1859582357-project-member] Acquired lock "refresh_cache-498c8e1e-18c6-4e09-a90d-106600bed2cd" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 646.356355] env[63088]: DEBUG nova.network.neutron [None req-b1b885bf-cef9-4808-8f16-707ab04499b3 tempest-ServersTestManualDisk-1859582357 tempest-ServersTestManualDisk-1859582357-project-member] [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 646.728877] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bb8843aa-5c19-4289-a41c-17cbfcb47a00 tempest-ServerActionsTestOtherA-1614359098 tempest-ServerActionsTestOtherA-1614359098-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 646.890295] env[63088]: DEBUG nova.network.neutron [None req-b1b885bf-cef9-4808-8f16-707ab04499b3 tempest-ServersTestManualDisk-1859582357 tempest-ServersTestManualDisk-1859582357-project-member] [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 647.062125] env[63088]: DEBUG nova.network.neutron [None req-b1b885bf-cef9-4808-8f16-707ab04499b3 tempest-ServersTestManualDisk-1859582357 tempest-ServersTestManualDisk-1859582357-project-member] [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 647.320707] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d90644d-adf0-479a-8e00-d26215ab8fa9 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.329650] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53265675-64e0-45b6-9420-28864098375c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.377120] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45035190-0c1e-49bb-983c-0803fb3167ec {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.386983] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28fb7ee1-2076-4bdd-b73d-120a7506bbff {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.401170] env[63088]: DEBUG nova.compute.provider_tree [None req-0564aade-729d-4cf1-a59d-513384386c54 tempest-AttachInterfacesV270Test-1783457293 tempest-AttachInterfacesV270Test-1783457293-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 647.563674] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b1b885bf-cef9-4808-8f16-707ab04499b3 tempest-ServersTestManualDisk-1859582357 tempest-ServersTestManualDisk-1859582357-project-member] Releasing lock "refresh_cache-498c8e1e-18c6-4e09-a90d-106600bed2cd" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 647.563974] env[63088]: DEBUG nova.compute.manager [None req-b1b885bf-cef9-4808-8f16-707ab04499b3 tempest-ServersTestManualDisk-1859582357 tempest-ServersTestManualDisk-1859582357-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 647.564328] env[63088]: DEBUG nova.compute.manager [None req-b1b885bf-cef9-4808-8f16-707ab04499b3 tempest-ServersTestManualDisk-1859582357 tempest-ServersTestManualDisk-1859582357-project-member] [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 647.564376] env[63088]: DEBUG nova.network.neutron [None req-b1b885bf-cef9-4808-8f16-707ab04499b3 tempest-ServersTestManualDisk-1859582357 tempest-ServersTestManualDisk-1859582357-project-member] [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 647.580564] env[63088]: DEBUG nova.network.neutron [None req-b1b885bf-cef9-4808-8f16-707ab04499b3 tempest-ServersTestManualDisk-1859582357 tempest-ServersTestManualDisk-1859582357-project-member] [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 647.731246] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3a58af89-5e90-4bc8-9f8f-c02b80d14311 tempest-InstanceActionsTestJSON-1717540784 tempest-InstanceActionsTestJSON-1717540784-project-member] Acquiring lock "7b5cc7b5-98b9-4344-b714-270dca798799" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 647.731246] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3a58af89-5e90-4bc8-9f8f-c02b80d14311 tempest-InstanceActionsTestJSON-1717540784 tempest-InstanceActionsTestJSON-1717540784-project-member] Lock "7b5cc7b5-98b9-4344-b714-270dca798799" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 647.906975] env[63088]: DEBUG nova.scheduler.client.report [None req-0564aade-729d-4cf1-a59d-513384386c54 tempest-AttachInterfacesV270Test-1783457293 tempest-AttachInterfacesV270Test-1783457293-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 648.057314] env[63088]: DEBUG nova.compute.manager [req-0a9caee3-2f27-41db-904c-f3fd8c3c0f99 req-4e986354-f028-4d1b-9819-90e4ae0186f1 service nova] [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] Received event network-changed-3aed79cc-3d41-4762-992b-05bb123fc735 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 648.057314] env[63088]: DEBUG nova.compute.manager [req-0a9caee3-2f27-41db-904c-f3fd8c3c0f99 req-4e986354-f028-4d1b-9819-90e4ae0186f1 service nova] [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] Refreshing instance network info cache due to event network-changed-3aed79cc-3d41-4762-992b-05bb123fc735. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 648.057314] env[63088]: DEBUG oslo_concurrency.lockutils [req-0a9caee3-2f27-41db-904c-f3fd8c3c0f99 req-4e986354-f028-4d1b-9819-90e4ae0186f1 service nova] Acquiring lock "refresh_cache-966fe303-09b9-449f-8bda-4f6cda135f7a" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 648.057314] env[63088]: DEBUG oslo_concurrency.lockutils [req-0a9caee3-2f27-41db-904c-f3fd8c3c0f99 req-4e986354-f028-4d1b-9819-90e4ae0186f1 service nova] Acquired lock "refresh_cache-966fe303-09b9-449f-8bda-4f6cda135f7a" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 648.057314] env[63088]: DEBUG nova.network.neutron [req-0a9caee3-2f27-41db-904c-f3fd8c3c0f99 req-4e986354-f028-4d1b-9819-90e4ae0186f1 service nova] [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] Refreshing network info cache for port 3aed79cc-3d41-4762-992b-05bb123fc735 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 648.087405] env[63088]: DEBUG nova.network.neutron [None req-b1b885bf-cef9-4808-8f16-707ab04499b3 tempest-ServersTestManualDisk-1859582357 tempest-ServersTestManualDisk-1859582357-project-member] [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 648.253924] env[63088]: ERROR nova.compute.manager [None req-a5e283fa-bcc1-4186-a4e3-1ae9f8d221af tempest-ServerMetadataTestJSON-859565163 tempest-ServerMetadataTestJSON-859565163-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3aed79cc-3d41-4762-992b-05bb123fc735, please check neutron logs for more information. [ 648.253924] env[63088]: ERROR nova.compute.manager Traceback (most recent call last): [ 648.253924] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 648.253924] env[63088]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 648.253924] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 648.253924] env[63088]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 648.253924] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 648.253924] env[63088]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 648.253924] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 648.253924] env[63088]: ERROR nova.compute.manager self.force_reraise() [ 648.253924] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 648.253924] env[63088]: ERROR nova.compute.manager raise self.value [ 648.253924] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 648.253924] env[63088]: ERROR nova.compute.manager updated_port = self._update_port( [ 648.253924] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 648.253924] env[63088]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 648.254950] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 648.254950] env[63088]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 648.254950] env[63088]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3aed79cc-3d41-4762-992b-05bb123fc735, please check neutron logs for more information. [ 648.254950] env[63088]: ERROR nova.compute.manager [ 648.254950] env[63088]: Traceback (most recent call last): [ 648.254950] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 648.254950] env[63088]: listener.cb(fileno) [ 648.254950] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 648.254950] env[63088]: result = function(*args, **kwargs) [ 648.254950] env[63088]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 648.254950] env[63088]: return func(*args, **kwargs) [ 648.254950] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 648.254950] env[63088]: raise e [ 648.254950] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 648.254950] env[63088]: nwinfo = self.network_api.allocate_for_instance( [ 648.254950] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 648.254950] env[63088]: created_port_ids = self._update_ports_for_instance( [ 648.254950] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 648.254950] env[63088]: with excutils.save_and_reraise_exception(): [ 648.254950] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 648.254950] env[63088]: self.force_reraise() [ 648.254950] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 648.254950] env[63088]: raise self.value [ 648.254950] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 648.254950] env[63088]: updated_port = self._update_port( [ 648.254950] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 648.254950] env[63088]: _ensure_no_port_binding_failure(port) [ 648.254950] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 648.254950] env[63088]: raise exception.PortBindingFailed(port_id=port['id']) [ 648.255799] env[63088]: nova.exception.PortBindingFailed: Binding failed for port 3aed79cc-3d41-4762-992b-05bb123fc735, please check neutron logs for more information. [ 648.255799] env[63088]: Removing descriptor: 15 [ 648.255799] env[63088]: ERROR nova.compute.manager [None req-a5e283fa-bcc1-4186-a4e3-1ae9f8d221af tempest-ServerMetadataTestJSON-859565163 tempest-ServerMetadataTestJSON-859565163-project-member] [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3aed79cc-3d41-4762-992b-05bb123fc735, please check neutron logs for more information. [ 648.255799] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] Traceback (most recent call last): [ 648.255799] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 648.255799] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] yield resources [ 648.255799] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 648.255799] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] self.driver.spawn(context, instance, image_meta, [ 648.255799] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 648.255799] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 648.255799] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 648.255799] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] vm_ref = self.build_virtual_machine(instance, [ 648.256110] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 648.256110] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] vif_infos = vmwarevif.get_vif_info(self._session, [ 648.256110] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 648.256110] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] for vif in network_info: [ 648.256110] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 648.256110] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] return self._sync_wrapper(fn, *args, **kwargs) [ 648.256110] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 648.256110] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] self.wait() [ 648.256110] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 648.256110] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] self[:] = self._gt.wait() [ 648.256110] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 648.256110] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] return self._exit_event.wait() [ 648.256110] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 648.256423] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] result = hub.switch() [ 648.256423] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 648.256423] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] return self.greenlet.switch() [ 648.256423] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 648.256423] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] result = function(*args, **kwargs) [ 648.256423] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 648.256423] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] return func(*args, **kwargs) [ 648.256423] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 648.256423] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] raise e [ 648.256423] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 648.256423] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] nwinfo = self.network_api.allocate_for_instance( [ 648.256423] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 648.256423] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] created_port_ids = self._update_ports_for_instance( [ 648.256754] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 648.256754] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] with excutils.save_and_reraise_exception(): [ 648.256754] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 648.256754] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] self.force_reraise() [ 648.256754] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 648.256754] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] raise self.value [ 648.256754] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 648.256754] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] updated_port = self._update_port( [ 648.256754] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 648.256754] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] _ensure_no_port_binding_failure(port) [ 648.256754] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 648.256754] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] raise exception.PortBindingFailed(port_id=port['id']) [ 648.257050] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] nova.exception.PortBindingFailed: Binding failed for port 3aed79cc-3d41-4762-992b-05bb123fc735, please check neutron logs for more information. [ 648.257050] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] [ 648.257050] env[63088]: INFO nova.compute.manager [None req-a5e283fa-bcc1-4186-a4e3-1ae9f8d221af tempest-ServerMetadataTestJSON-859565163 tempest-ServerMetadataTestJSON-859565163-project-member] [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] Terminating instance [ 648.257994] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a5e283fa-bcc1-4186-a4e3-1ae9f8d221af tempest-ServerMetadataTestJSON-859565163 tempest-ServerMetadataTestJSON-859565163-project-member] Acquiring lock "refresh_cache-966fe303-09b9-449f-8bda-4f6cda135f7a" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 648.413354] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0564aade-729d-4cf1-a59d-513384386c54 tempest-AttachInterfacesV270Test-1783457293 tempest-AttachInterfacesV270Test-1783457293-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.062s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 648.413992] env[63088]: ERROR nova.compute.manager [None req-0564aade-729d-4cf1-a59d-513384386c54 tempest-AttachInterfacesV270Test-1783457293 tempest-AttachInterfacesV270Test-1783457293-project-member] [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c7b197ee-fb06-46ec-9aac-b7fb7874113c, please check neutron logs for more information. [ 648.413992] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] Traceback (most recent call last): [ 648.413992] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 648.413992] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] self.driver.spawn(context, instance, image_meta, [ 648.413992] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 648.413992] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] self._vmops.spawn(context, instance, image_meta, injected_files, [ 648.413992] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 648.413992] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] vm_ref = self.build_virtual_machine(instance, [ 648.413992] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 648.413992] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] vif_infos = vmwarevif.get_vif_info(self._session, [ 648.413992] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 648.414339] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] for vif in network_info: [ 648.414339] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 648.414339] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] return self._sync_wrapper(fn, *args, **kwargs) [ 648.414339] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 648.414339] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] self.wait() [ 648.414339] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 648.414339] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] self[:] = self._gt.wait() [ 648.414339] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 648.414339] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] return self._exit_event.wait() [ 648.414339] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 648.414339] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] result = hub.switch() [ 648.414339] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 648.414339] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] return self.greenlet.switch() [ 648.414630] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 648.414630] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] result = function(*args, **kwargs) [ 648.414630] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 648.414630] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] return func(*args, **kwargs) [ 648.414630] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 648.414630] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] raise e [ 648.414630] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 648.414630] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] nwinfo = self.network_api.allocate_for_instance( [ 648.414630] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 648.414630] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] created_port_ids = self._update_ports_for_instance( [ 648.414630] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 648.414630] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] with excutils.save_and_reraise_exception(): [ 648.414630] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 648.414930] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] self.force_reraise() [ 648.414930] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 648.414930] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] raise self.value [ 648.414930] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 648.414930] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] updated_port = self._update_port( [ 648.414930] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 648.414930] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] _ensure_no_port_binding_failure(port) [ 648.414930] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 648.414930] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] raise exception.PortBindingFailed(port_id=port['id']) [ 648.414930] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] nova.exception.PortBindingFailed: Binding failed for port c7b197ee-fb06-46ec-9aac-b7fb7874113c, please check neutron logs for more information. [ 648.414930] env[63088]: ERROR nova.compute.manager [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] [ 648.415210] env[63088]: DEBUG nova.compute.utils [None req-0564aade-729d-4cf1-a59d-513384386c54 tempest-AttachInterfacesV270Test-1783457293 tempest-AttachInterfacesV270Test-1783457293-project-member] [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] Binding failed for port c7b197ee-fb06-46ec-9aac-b7fb7874113c, please check neutron logs for more information. {{(pid=63088) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 648.416466] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f81b332d-72eb-4149-a6a2-3c8a9c9bec53 tempest-ServersNegativeTestJSON-1008333314 tempest-ServersNegativeTestJSON-1008333314-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.309s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 648.424039] env[63088]: DEBUG nova.compute.manager [None req-0564aade-729d-4cf1-a59d-513384386c54 tempest-AttachInterfacesV270Test-1783457293 tempest-AttachInterfacesV270Test-1783457293-project-member] [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] Build of instance 421e0f76-024d-4793-8f00-98731bb6b1af was re-scheduled: Binding failed for port c7b197ee-fb06-46ec-9aac-b7fb7874113c, please check neutron logs for more information. {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 648.424039] env[63088]: DEBUG nova.compute.manager [None req-0564aade-729d-4cf1-a59d-513384386c54 tempest-AttachInterfacesV270Test-1783457293 tempest-AttachInterfacesV270Test-1783457293-project-member] [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] Unplugging VIFs for instance {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 648.424039] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0564aade-729d-4cf1-a59d-513384386c54 tempest-AttachInterfacesV270Test-1783457293 tempest-AttachInterfacesV270Test-1783457293-project-member] Acquiring lock "refresh_cache-421e0f76-024d-4793-8f00-98731bb6b1af" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 648.424039] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0564aade-729d-4cf1-a59d-513384386c54 tempest-AttachInterfacesV270Test-1783457293 tempest-AttachInterfacesV270Test-1783457293-project-member] Acquired lock "refresh_cache-421e0f76-024d-4793-8f00-98731bb6b1af" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 648.424270] env[63088]: DEBUG nova.network.neutron [None req-0564aade-729d-4cf1-a59d-513384386c54 tempest-AttachInterfacesV270Test-1783457293 tempest-AttachInterfacesV270Test-1783457293-project-member] [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 648.585065] env[63088]: DEBUG nova.network.neutron [req-0a9caee3-2f27-41db-904c-f3fd8c3c0f99 req-4e986354-f028-4d1b-9819-90e4ae0186f1 service nova] [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 648.588560] env[63088]: INFO nova.compute.manager [None req-b1b885bf-cef9-4808-8f16-707ab04499b3 tempest-ServersTestManualDisk-1859582357 tempest-ServersTestManualDisk-1859582357-project-member] [instance: 498c8e1e-18c6-4e09-a90d-106600bed2cd] Took 1.02 seconds to deallocate network for instance. [ 648.703158] env[63088]: DEBUG nova.network.neutron [req-0a9caee3-2f27-41db-904c-f3fd8c3c0f99 req-4e986354-f028-4d1b-9819-90e4ae0186f1 service nova] [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 648.958017] env[63088]: DEBUG nova.network.neutron [None req-0564aade-729d-4cf1-a59d-513384386c54 tempest-AttachInterfacesV270Test-1783457293 tempest-AttachInterfacesV270Test-1783457293-project-member] [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 649.097843] env[63088]: DEBUG nova.network.neutron [None req-0564aade-729d-4cf1-a59d-513384386c54 tempest-AttachInterfacesV270Test-1783457293 tempest-AttachInterfacesV270Test-1783457293-project-member] [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 649.205997] env[63088]: DEBUG oslo_concurrency.lockutils [req-0a9caee3-2f27-41db-904c-f3fd8c3c0f99 req-4e986354-f028-4d1b-9819-90e4ae0186f1 service nova] Releasing lock "refresh_cache-966fe303-09b9-449f-8bda-4f6cda135f7a" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 649.206446] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a5e283fa-bcc1-4186-a4e3-1ae9f8d221af tempest-ServerMetadataTestJSON-859565163 tempest-ServerMetadataTestJSON-859565163-project-member] Acquired lock "refresh_cache-966fe303-09b9-449f-8bda-4f6cda135f7a" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 649.206977] env[63088]: DEBUG nova.network.neutron [None req-a5e283fa-bcc1-4186-a4e3-1ae9f8d221af tempest-ServerMetadataTestJSON-859565163 tempest-ServerMetadataTestJSON-859565163-project-member] [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 649.405083] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a574dc1-0094-4b02-b713-42c21d99b990 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.415019] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-636522f0-4d1f-4e7c-9637-b86a0ef0fb3d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.449314] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf45e1a3-8d5e-46d9-bca9-ba03cdb7c1f2 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.459632] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83b3689e-4bcf-4d8e-b5c0-abaac2398bfc {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.478556] env[63088]: DEBUG nova.compute.provider_tree [None req-f81b332d-72eb-4149-a6a2-3c8a9c9bec53 tempest-ServersNegativeTestJSON-1008333314 tempest-ServersNegativeTestJSON-1008333314-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 649.602782] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0564aade-729d-4cf1-a59d-513384386c54 tempest-AttachInterfacesV270Test-1783457293 tempest-AttachInterfacesV270Test-1783457293-project-member] Releasing lock "refresh_cache-421e0f76-024d-4793-8f00-98731bb6b1af" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 649.602782] env[63088]: DEBUG nova.compute.manager [None req-0564aade-729d-4cf1-a59d-513384386c54 tempest-AttachInterfacesV270Test-1783457293 tempest-AttachInterfacesV270Test-1783457293-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 649.602961] env[63088]: DEBUG nova.compute.manager [None req-0564aade-729d-4cf1-a59d-513384386c54 tempest-AttachInterfacesV270Test-1783457293 tempest-AttachInterfacesV270Test-1783457293-project-member] [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 649.603159] env[63088]: DEBUG nova.network.neutron [None req-0564aade-729d-4cf1-a59d-513384386c54 tempest-AttachInterfacesV270Test-1783457293 tempest-AttachInterfacesV270Test-1783457293-project-member] [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 649.621097] env[63088]: DEBUG nova.network.neutron [None req-0564aade-729d-4cf1-a59d-513384386c54 tempest-AttachInterfacesV270Test-1783457293 tempest-AttachInterfacesV270Test-1783457293-project-member] [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 649.628276] env[63088]: INFO nova.scheduler.client.report [None req-b1b885bf-cef9-4808-8f16-707ab04499b3 tempest-ServersTestManualDisk-1859582357 tempest-ServersTestManualDisk-1859582357-project-member] Deleted allocations for instance 498c8e1e-18c6-4e09-a90d-106600bed2cd [ 649.727895] env[63088]: DEBUG nova.network.neutron [None req-a5e283fa-bcc1-4186-a4e3-1ae9f8d221af tempest-ServerMetadataTestJSON-859565163 tempest-ServerMetadataTestJSON-859565163-project-member] [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 649.807893] env[63088]: DEBUG nova.network.neutron [None req-a5e283fa-bcc1-4186-a4e3-1ae9f8d221af tempest-ServerMetadataTestJSON-859565163 tempest-ServerMetadataTestJSON-859565163-project-member] [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 649.983110] env[63088]: DEBUG nova.scheduler.client.report [None req-f81b332d-72eb-4149-a6a2-3c8a9c9bec53 tempest-ServersNegativeTestJSON-1008333314 tempest-ServersNegativeTestJSON-1008333314-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 650.073335] env[63088]: DEBUG oslo_concurrency.lockutils [None req-87ab5d07-9ebb-429d-8640-0197960e7d1a tempest-ServersNegativeTestMultiTenantJSON-2054426678 tempest-ServersNegativeTestMultiTenantJSON-2054426678-project-member] Acquiring lock "d851a26b-48aa-4ffd-9c23-cea27543e0fa" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 650.073571] env[63088]: DEBUG oslo_concurrency.lockutils [None req-87ab5d07-9ebb-429d-8640-0197960e7d1a tempest-ServersNegativeTestMultiTenantJSON-2054426678 tempest-ServersNegativeTestMultiTenantJSON-2054426678-project-member] Lock "d851a26b-48aa-4ffd-9c23-cea27543e0fa" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 650.123647] env[63088]: DEBUG nova.network.neutron [None req-0564aade-729d-4cf1-a59d-513384386c54 tempest-AttachInterfacesV270Test-1783457293 tempest-AttachInterfacesV270Test-1783457293-project-member] [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 650.141783] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b1b885bf-cef9-4808-8f16-707ab04499b3 tempest-ServersTestManualDisk-1859582357 tempest-ServersTestManualDisk-1859582357-project-member] Lock "498c8e1e-18c6-4e09-a90d-106600bed2cd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 75.509s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 650.225219] env[63088]: DEBUG nova.compute.manager [req-ddc8ebd6-28b5-44ab-99e4-28b14a409d4e req-f2833d37-1129-4cf9-8f6c-d23ce70aa2a5 service nova] [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] Received event network-vif-deleted-3aed79cc-3d41-4762-992b-05bb123fc735 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 650.310889] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a5e283fa-bcc1-4186-a4e3-1ae9f8d221af tempest-ServerMetadataTestJSON-859565163 tempest-ServerMetadataTestJSON-859565163-project-member] Releasing lock "refresh_cache-966fe303-09b9-449f-8bda-4f6cda135f7a" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 650.311439] env[63088]: DEBUG nova.compute.manager [None req-a5e283fa-bcc1-4186-a4e3-1ae9f8d221af tempest-ServerMetadataTestJSON-859565163 tempest-ServerMetadataTestJSON-859565163-project-member] [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 650.311664] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-a5e283fa-bcc1-4186-a4e3-1ae9f8d221af tempest-ServerMetadataTestJSON-859565163 tempest-ServerMetadataTestJSON-859565163-project-member] [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 650.311950] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f704a353-ed28-4956-a86b-4c5e2698d76b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.322166] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1fa9bf9-8ffc-4871-a49a-216492c122e0 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.350104] env[63088]: WARNING nova.virt.vmwareapi.vmops [None req-a5e283fa-bcc1-4186-a4e3-1ae9f8d221af tempest-ServerMetadataTestJSON-859565163 tempest-ServerMetadataTestJSON-859565163-project-member] [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 966fe303-09b9-449f-8bda-4f6cda135f7a could not be found. [ 650.350104] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-a5e283fa-bcc1-4186-a4e3-1ae9f8d221af tempest-ServerMetadataTestJSON-859565163 tempest-ServerMetadataTestJSON-859565163-project-member] [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 650.350104] env[63088]: INFO nova.compute.manager [None req-a5e283fa-bcc1-4186-a4e3-1ae9f8d221af tempest-ServerMetadataTestJSON-859565163 tempest-ServerMetadataTestJSON-859565163-project-member] [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] Took 0.04 seconds to destroy the instance on the hypervisor. [ 650.350104] env[63088]: DEBUG oslo.service.loopingcall [None req-a5e283fa-bcc1-4186-a4e3-1ae9f8d221af tempest-ServerMetadataTestJSON-859565163 tempest-ServerMetadataTestJSON-859565163-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 650.350104] env[63088]: DEBUG nova.compute.manager [-] [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 650.350104] env[63088]: DEBUG nova.network.neutron [-] [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 650.365037] env[63088]: DEBUG nova.network.neutron [-] [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 650.488766] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f81b332d-72eb-4149-a6a2-3c8a9c9bec53 tempest-ServersNegativeTestJSON-1008333314 tempest-ServersNegativeTestJSON-1008333314-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.072s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 650.489531] env[63088]: ERROR nova.compute.manager [None req-f81b332d-72eb-4149-a6a2-3c8a9c9bec53 tempest-ServersNegativeTestJSON-1008333314 tempest-ServersNegativeTestJSON-1008333314-project-member] [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a0fc3b2d-a2f9-457b-acc2-16a9ac49a5bd, please check neutron logs for more information. [ 650.489531] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] Traceback (most recent call last): [ 650.489531] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 650.489531] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] self.driver.spawn(context, instance, image_meta, [ 650.489531] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 650.489531] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 650.489531] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 650.489531] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] vm_ref = self.build_virtual_machine(instance, [ 650.489531] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 650.489531] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] vif_infos = vmwarevif.get_vif_info(self._session, [ 650.489531] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 650.490109] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] for vif in network_info: [ 650.490109] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 650.490109] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] return self._sync_wrapper(fn, *args, **kwargs) [ 650.490109] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 650.490109] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] self.wait() [ 650.490109] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 650.490109] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] self[:] = self._gt.wait() [ 650.490109] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 650.490109] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] return self._exit_event.wait() [ 650.490109] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 650.490109] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] result = hub.switch() [ 650.490109] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 650.490109] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] return self.greenlet.switch() [ 650.490968] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 650.490968] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] result = function(*args, **kwargs) [ 650.490968] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 650.490968] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] return func(*args, **kwargs) [ 650.490968] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 650.490968] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] raise e [ 650.490968] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 650.490968] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] nwinfo = self.network_api.allocate_for_instance( [ 650.490968] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 650.490968] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] created_port_ids = self._update_ports_for_instance( [ 650.490968] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 650.490968] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] with excutils.save_and_reraise_exception(): [ 650.490968] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 650.491613] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] self.force_reraise() [ 650.491613] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 650.491613] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] raise self.value [ 650.491613] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 650.491613] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] updated_port = self._update_port( [ 650.491613] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 650.491613] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] _ensure_no_port_binding_failure(port) [ 650.491613] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 650.491613] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] raise exception.PortBindingFailed(port_id=port['id']) [ 650.491613] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] nova.exception.PortBindingFailed: Binding failed for port a0fc3b2d-a2f9-457b-acc2-16a9ac49a5bd, please check neutron logs for more information. [ 650.491613] env[63088]: ERROR nova.compute.manager [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] [ 650.491935] env[63088]: DEBUG nova.compute.utils [None req-f81b332d-72eb-4149-a6a2-3c8a9c9bec53 tempest-ServersNegativeTestJSON-1008333314 tempest-ServersNegativeTestJSON-1008333314-project-member] [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] Binding failed for port a0fc3b2d-a2f9-457b-acc2-16a9ac49a5bd, please check neutron logs for more information. {{(pid=63088) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 650.491935] env[63088]: DEBUG oslo_concurrency.lockutils [None req-224c474a-3c21-4a3f-9526-5bb4e3ba7d78 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.783s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 650.493773] env[63088]: INFO nova.compute.claims [None req-224c474a-3c21-4a3f-9526-5bb4e3ba7d78 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 650.498087] env[63088]: DEBUG nova.compute.manager [None req-f81b332d-72eb-4149-a6a2-3c8a9c9bec53 tempest-ServersNegativeTestJSON-1008333314 tempest-ServersNegativeTestJSON-1008333314-project-member] [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] Build of instance 5b778247-1434-447e-8ff1-7359bb9e8ea2 was re-scheduled: Binding failed for port a0fc3b2d-a2f9-457b-acc2-16a9ac49a5bd, please check neutron logs for more information. {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 650.498729] env[63088]: DEBUG nova.compute.manager [None req-f81b332d-72eb-4149-a6a2-3c8a9c9bec53 tempest-ServersNegativeTestJSON-1008333314 tempest-ServersNegativeTestJSON-1008333314-project-member] [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] Unplugging VIFs for instance {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 650.500085] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f81b332d-72eb-4149-a6a2-3c8a9c9bec53 tempest-ServersNegativeTestJSON-1008333314 tempest-ServersNegativeTestJSON-1008333314-project-member] Acquiring lock "refresh_cache-5b778247-1434-447e-8ff1-7359bb9e8ea2" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 650.500085] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f81b332d-72eb-4149-a6a2-3c8a9c9bec53 tempest-ServersNegativeTestJSON-1008333314 tempest-ServersNegativeTestJSON-1008333314-project-member] Acquired lock "refresh_cache-5b778247-1434-447e-8ff1-7359bb9e8ea2" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 650.500085] env[63088]: DEBUG nova.network.neutron [None req-f81b332d-72eb-4149-a6a2-3c8a9c9bec53 tempest-ServersNegativeTestJSON-1008333314 tempest-ServersNegativeTestJSON-1008333314-project-member] [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 650.631654] env[63088]: INFO nova.compute.manager [None req-0564aade-729d-4cf1-a59d-513384386c54 tempest-AttachInterfacesV270Test-1783457293 tempest-AttachInterfacesV270Test-1783457293-project-member] [instance: 421e0f76-024d-4793-8f00-98731bb6b1af] Took 1.03 seconds to deallocate network for instance. [ 650.644878] env[63088]: DEBUG nova.compute.manager [None req-18116a16-f5bf-4db6-aaca-b64d4351d5d4 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 650.868139] env[63088]: DEBUG nova.network.neutron [-] [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 650.927720] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 650.927937] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 651.026397] env[63088]: DEBUG nova.network.neutron [None req-f81b332d-72eb-4149-a6a2-3c8a9c9bec53 tempest-ServersNegativeTestJSON-1008333314 tempest-ServersNegativeTestJSON-1008333314-project-member] [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 651.146157] env[63088]: DEBUG nova.network.neutron [None req-f81b332d-72eb-4149-a6a2-3c8a9c9bec53 tempest-ServersNegativeTestJSON-1008333314 tempest-ServersNegativeTestJSON-1008333314-project-member] [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 651.177181] env[63088]: DEBUG oslo_concurrency.lockutils [None req-18116a16-f5bf-4db6-aaca-b64d4351d5d4 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 651.375688] env[63088]: INFO nova.compute.manager [-] [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] Took 1.03 seconds to deallocate network for instance. [ 651.377390] env[63088]: DEBUG nova.compute.claims [None req-a5e283fa-bcc1-4186-a4e3-1ae9f8d221af tempest-ServerMetadataTestJSON-859565163 tempest-ServerMetadataTestJSON-859565163-project-member] [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] Aborting claim: {{(pid=63088) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 651.377390] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a5e283fa-bcc1-4186-a4e3-1ae9f8d221af tempest-ServerMetadataTestJSON-859565163 tempest-ServerMetadataTestJSON-859565163-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 651.432173] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 651.435589] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Starting heal instance info cache {{(pid=63088) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 651.435589] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Rebuilding the list of instances to heal {{(pid=63088) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 651.649766] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f81b332d-72eb-4149-a6a2-3c8a9c9bec53 tempest-ServersNegativeTestJSON-1008333314 tempest-ServersNegativeTestJSON-1008333314-project-member] Releasing lock "refresh_cache-5b778247-1434-447e-8ff1-7359bb9e8ea2" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 651.650062] env[63088]: DEBUG nova.compute.manager [None req-f81b332d-72eb-4149-a6a2-3c8a9c9bec53 tempest-ServersNegativeTestJSON-1008333314 tempest-ServersNegativeTestJSON-1008333314-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 651.650258] env[63088]: DEBUG nova.compute.manager [None req-f81b332d-72eb-4149-a6a2-3c8a9c9bec53 tempest-ServersNegativeTestJSON-1008333314 tempest-ServersNegativeTestJSON-1008333314-project-member] [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 651.650441] env[63088]: DEBUG nova.network.neutron [None req-f81b332d-72eb-4149-a6a2-3c8a9c9bec53 tempest-ServersNegativeTestJSON-1008333314 tempest-ServersNegativeTestJSON-1008333314-project-member] [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 651.670016] env[63088]: INFO nova.scheduler.client.report [None req-0564aade-729d-4cf1-a59d-513384386c54 tempest-AttachInterfacesV270Test-1783457293 tempest-AttachInterfacesV270Test-1783457293-project-member] Deleted allocations for instance 421e0f76-024d-4793-8f00-98731bb6b1af [ 651.685298] env[63088]: DEBUG nova.network.neutron [None req-f81b332d-72eb-4149-a6a2-3c8a9c9bec53 tempest-ServersNegativeTestJSON-1008333314 tempest-ServersNegativeTestJSON-1008333314-project-member] [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 651.910868] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5ab12f0-3c80-4e9f-b94b-1f6beb325829 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.920787] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c09368e7-95b9-400a-8c23-7bb1b12dc228 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.960224] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] Skipping network cache update for instance because it is Building. {{(pid=63088) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 651.960534] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] Skipping network cache update for instance because it is Building. {{(pid=63088) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 651.960765] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Didn't find any instances for network info cache update. {{(pid=63088) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10012}} [ 651.961277] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 651.962382] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fce3b3c0-258f-4820-a26f-dc5954667a23 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.964849] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 651.965061] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 651.965695] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 651.965948] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 651.966107] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 651.966249] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63088) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 651.966400] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager.update_available_resource {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 651.975741] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1355cf83-4ca9-4cd9-83df-740c3f76d850 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.995298] env[63088]: DEBUG nova.compute.provider_tree [None req-224c474a-3c21-4a3f-9526-5bb4e3ba7d78 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 652.185483] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0564aade-729d-4cf1-a59d-513384386c54 tempest-AttachInterfacesV270Test-1783457293 tempest-AttachInterfacesV270Test-1783457293-project-member] Lock "421e0f76-024d-4793-8f00-98731bb6b1af" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 76.543s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 652.188148] env[63088]: DEBUG nova.network.neutron [None req-f81b332d-72eb-4149-a6a2-3c8a9c9bec53 tempest-ServersNegativeTestJSON-1008333314 tempest-ServersNegativeTestJSON-1008333314-project-member] [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 652.341229] env[63088]: DEBUG oslo_concurrency.lockutils [None req-aa39901a-ab16-4a54-a2f6-147052a0ed90 tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] Acquiring lock "43ded929-6890-4296-ae90-230f7c8b2595" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 652.341229] env[63088]: DEBUG oslo_concurrency.lockutils [None req-aa39901a-ab16-4a54-a2f6-147052a0ed90 tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] Lock "43ded929-6890-4296-ae90-230f7c8b2595" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 652.471057] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 652.499108] env[63088]: DEBUG nova.scheduler.client.report [None req-224c474a-3c21-4a3f-9526-5bb4e3ba7d78 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 652.692759] env[63088]: DEBUG nova.compute.manager [None req-f237740d-6eaf-4423-b770-dd9775fe6fd5 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 652.696279] env[63088]: INFO nova.compute.manager [None req-f81b332d-72eb-4149-a6a2-3c8a9c9bec53 tempest-ServersNegativeTestJSON-1008333314 tempest-ServersNegativeTestJSON-1008333314-project-member] [instance: 5b778247-1434-447e-8ff1-7359bb9e8ea2] Took 1.04 seconds to deallocate network for instance. [ 653.005139] env[63088]: DEBUG oslo_concurrency.lockutils [None req-224c474a-3c21-4a3f-9526-5bb4e3ba7d78 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.512s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 653.005702] env[63088]: DEBUG nova.compute.manager [None req-224c474a-3c21-4a3f-9526-5bb4e3ba7d78 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 653.009477] env[63088]: DEBUG oslo_concurrency.lockutils [None req-cab9e1ae-5fd3-4816-8684-8d11c94f36d2 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.387s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 653.011605] env[63088]: INFO nova.compute.claims [None req-cab9e1ae-5fd3-4816-8684-8d11c94f36d2 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 653.231517] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f237740d-6eaf-4423-b770-dd9775fe6fd5 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 653.519488] env[63088]: DEBUG nova.compute.utils [None req-224c474a-3c21-4a3f-9526-5bb4e3ba7d78 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 653.520233] env[63088]: DEBUG nova.compute.manager [None req-224c474a-3c21-4a3f-9526-5bb4e3ba7d78 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 653.520539] env[63088]: DEBUG nova.network.neutron [None req-224c474a-3c21-4a3f-9526-5bb4e3ba7d78 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 653.638283] env[63088]: DEBUG nova.policy [None req-224c474a-3c21-4a3f-9526-5bb4e3ba7d78 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ba9caaadce804132b4df2bc441099f06', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd593db530697467cad3671cc591a117e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 653.736712] env[63088]: INFO nova.scheduler.client.report [None req-f81b332d-72eb-4149-a6a2-3c8a9c9bec53 tempest-ServersNegativeTestJSON-1008333314 tempest-ServersNegativeTestJSON-1008333314-project-member] Deleted allocations for instance 5b778247-1434-447e-8ff1-7359bb9e8ea2 [ 654.026632] env[63088]: DEBUG nova.compute.manager [None req-224c474a-3c21-4a3f-9526-5bb4e3ba7d78 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 654.235158] env[63088]: DEBUG nova.network.neutron [None req-224c474a-3c21-4a3f-9526-5bb4e3ba7d78 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] Successfully created port: bfee4990-260b-40da-bbcf-869c85e6b40f {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 654.248408] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f81b332d-72eb-4149-a6a2-3c8a9c9bec53 tempest-ServersNegativeTestJSON-1008333314 tempest-ServersNegativeTestJSON-1008333314-project-member] Lock "5b778247-1434-447e-8ff1-7359bb9e8ea2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 74.856s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 654.500279] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b316e826-9021-4a0e-a912-4b1dc2f31f4f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.510338] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85d47ed8-1b13-4852-aad7-6e3bec19fb20 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.555068] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dab5bb4-910d-4125-a6ae-fdea14a91538 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.565011] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fac4cfd-b6fd-46cb-92c2-719fb2518b45 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.581098] env[63088]: DEBUG nova.compute.provider_tree [None req-cab9e1ae-5fd3-4816-8684-8d11c94f36d2 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 654.756609] env[63088]: DEBUG nova.compute.manager [None req-393edcfb-0595-4190-8ce0-4c14157dbbff tempest-ServerDiagnosticsNegativeTest-1508062492 tempest-ServerDiagnosticsNegativeTest-1508062492-project-member] [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 655.055853] env[63088]: DEBUG nova.compute.manager [None req-224c474a-3c21-4a3f-9526-5bb4e3ba7d78 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 655.084208] env[63088]: DEBUG nova.scheduler.client.report [None req-cab9e1ae-5fd3-4816-8684-8d11c94f36d2 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 655.102717] env[63088]: DEBUG nova.virt.hardware [None req-224c474a-3c21-4a3f-9526-5bb4e3ba7d78 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 655.103328] env[63088]: DEBUG nova.virt.hardware [None req-224c474a-3c21-4a3f-9526-5bb4e3ba7d78 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 655.103328] env[63088]: DEBUG nova.virt.hardware [None req-224c474a-3c21-4a3f-9526-5bb4e3ba7d78 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 655.103649] env[63088]: DEBUG nova.virt.hardware [None req-224c474a-3c21-4a3f-9526-5bb4e3ba7d78 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 655.103891] env[63088]: DEBUG nova.virt.hardware [None req-224c474a-3c21-4a3f-9526-5bb4e3ba7d78 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 655.104165] env[63088]: DEBUG nova.virt.hardware [None req-224c474a-3c21-4a3f-9526-5bb4e3ba7d78 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 655.104449] env[63088]: DEBUG nova.virt.hardware [None req-224c474a-3c21-4a3f-9526-5bb4e3ba7d78 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 655.105626] env[63088]: DEBUG nova.virt.hardware [None req-224c474a-3c21-4a3f-9526-5bb4e3ba7d78 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 655.105626] env[63088]: DEBUG nova.virt.hardware [None req-224c474a-3c21-4a3f-9526-5bb4e3ba7d78 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 655.105626] env[63088]: DEBUG nova.virt.hardware [None req-224c474a-3c21-4a3f-9526-5bb4e3ba7d78 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 655.105626] env[63088]: DEBUG nova.virt.hardware [None req-224c474a-3c21-4a3f-9526-5bb4e3ba7d78 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 655.106653] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fa8fb74-11ce-42ca-9e38-dc24487d1473 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.118958] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7ed8d35-0a37-4e2e-b35f-ad6c825edfd9 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.294023] env[63088]: DEBUG oslo_concurrency.lockutils [None req-393edcfb-0595-4190-8ce0-4c14157dbbff tempest-ServerDiagnosticsNegativeTest-1508062492 tempest-ServerDiagnosticsNegativeTest-1508062492-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 655.590306] env[63088]: DEBUG oslo_concurrency.lockutils [None req-cab9e1ae-5fd3-4816-8684-8d11c94f36d2 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.581s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 655.590832] env[63088]: DEBUG nova.compute.manager [None req-cab9e1ae-5fd3-4816-8684-8d11c94f36d2 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 655.597047] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b7ea81d3-fe92-4160-8c6b-de9afb157d35 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.948s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 655.597047] env[63088]: INFO nova.compute.claims [None req-b7ea81d3-fe92-4160-8c6b-de9afb157d35 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 655.844828] env[63088]: DEBUG nova.compute.manager [req-250e3fc5-7138-4664-9bc2-9ce905f618b8 req-d2ecd7dd-ecad-4143-99aa-10d4aa97215c service nova] [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] Received event network-changed-bfee4990-260b-40da-bbcf-869c85e6b40f {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 655.845131] env[63088]: DEBUG nova.compute.manager [req-250e3fc5-7138-4664-9bc2-9ce905f618b8 req-d2ecd7dd-ecad-4143-99aa-10d4aa97215c service nova] [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] Refreshing instance network info cache due to event network-changed-bfee4990-260b-40da-bbcf-869c85e6b40f. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 655.845422] env[63088]: DEBUG oslo_concurrency.lockutils [req-250e3fc5-7138-4664-9bc2-9ce905f618b8 req-d2ecd7dd-ecad-4143-99aa-10d4aa97215c service nova] Acquiring lock "refresh_cache-62bd4013-f007-4f69-b037-14bedf1e63cf" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 655.845621] env[63088]: DEBUG oslo_concurrency.lockutils [req-250e3fc5-7138-4664-9bc2-9ce905f618b8 req-d2ecd7dd-ecad-4143-99aa-10d4aa97215c service nova] Acquired lock "refresh_cache-62bd4013-f007-4f69-b037-14bedf1e63cf" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 655.845864] env[63088]: DEBUG nova.network.neutron [req-250e3fc5-7138-4664-9bc2-9ce905f618b8 req-d2ecd7dd-ecad-4143-99aa-10d4aa97215c service nova] [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] Refreshing network info cache for port bfee4990-260b-40da-bbcf-869c85e6b40f {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 656.001175] env[63088]: ERROR nova.compute.manager [None req-224c474a-3c21-4a3f-9526-5bb4e3ba7d78 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port bfee4990-260b-40da-bbcf-869c85e6b40f, please check neutron logs for more information. [ 656.001175] env[63088]: ERROR nova.compute.manager Traceback (most recent call last): [ 656.001175] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 656.001175] env[63088]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 656.001175] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 656.001175] env[63088]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 656.001175] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 656.001175] env[63088]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 656.001175] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 656.001175] env[63088]: ERROR nova.compute.manager self.force_reraise() [ 656.001175] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 656.001175] env[63088]: ERROR nova.compute.manager raise self.value [ 656.001175] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 656.001175] env[63088]: ERROR nova.compute.manager updated_port = self._update_port( [ 656.001175] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 656.001175] env[63088]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 656.001822] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 656.001822] env[63088]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 656.001822] env[63088]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port bfee4990-260b-40da-bbcf-869c85e6b40f, please check neutron logs for more information. [ 656.001822] env[63088]: ERROR nova.compute.manager [ 656.001822] env[63088]: Traceback (most recent call last): [ 656.001822] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 656.001822] env[63088]: listener.cb(fileno) [ 656.001822] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 656.001822] env[63088]: result = function(*args, **kwargs) [ 656.001822] env[63088]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 656.001822] env[63088]: return func(*args, **kwargs) [ 656.001822] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 656.001822] env[63088]: raise e [ 656.001822] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 656.001822] env[63088]: nwinfo = self.network_api.allocate_for_instance( [ 656.001822] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 656.001822] env[63088]: created_port_ids = self._update_ports_for_instance( [ 656.001822] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 656.001822] env[63088]: with excutils.save_and_reraise_exception(): [ 656.001822] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 656.001822] env[63088]: self.force_reraise() [ 656.001822] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 656.001822] env[63088]: raise self.value [ 656.001822] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 656.001822] env[63088]: updated_port = self._update_port( [ 656.001822] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 656.001822] env[63088]: _ensure_no_port_binding_failure(port) [ 656.001822] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 656.001822] env[63088]: raise exception.PortBindingFailed(port_id=port['id']) [ 656.002621] env[63088]: nova.exception.PortBindingFailed: Binding failed for port bfee4990-260b-40da-bbcf-869c85e6b40f, please check neutron logs for more information. [ 656.002621] env[63088]: Removing descriptor: 15 [ 656.002621] env[63088]: ERROR nova.compute.manager [None req-224c474a-3c21-4a3f-9526-5bb4e3ba7d78 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port bfee4990-260b-40da-bbcf-869c85e6b40f, please check neutron logs for more information. [ 656.002621] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] Traceback (most recent call last): [ 656.002621] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 656.002621] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] yield resources [ 656.002621] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 656.002621] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] self.driver.spawn(context, instance, image_meta, [ 656.002621] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 656.002621] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] self._vmops.spawn(context, instance, image_meta, injected_files, [ 656.002621] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 656.002621] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] vm_ref = self.build_virtual_machine(instance, [ 656.002965] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 656.002965] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] vif_infos = vmwarevif.get_vif_info(self._session, [ 656.002965] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 656.002965] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] for vif in network_info: [ 656.002965] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 656.002965] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] return self._sync_wrapper(fn, *args, **kwargs) [ 656.002965] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 656.002965] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] self.wait() [ 656.002965] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 656.002965] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] self[:] = self._gt.wait() [ 656.002965] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 656.002965] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] return self._exit_event.wait() [ 656.002965] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 656.003366] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] result = hub.switch() [ 656.003366] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 656.003366] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] return self.greenlet.switch() [ 656.003366] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 656.003366] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] result = function(*args, **kwargs) [ 656.003366] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 656.003366] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] return func(*args, **kwargs) [ 656.003366] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 656.003366] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] raise e [ 656.003366] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 656.003366] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] nwinfo = self.network_api.allocate_for_instance( [ 656.003366] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 656.003366] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] created_port_ids = self._update_ports_for_instance( [ 656.003809] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 656.003809] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] with excutils.save_and_reraise_exception(): [ 656.003809] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 656.003809] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] self.force_reraise() [ 656.003809] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 656.003809] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] raise self.value [ 656.003809] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 656.003809] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] updated_port = self._update_port( [ 656.003809] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 656.003809] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] _ensure_no_port_binding_failure(port) [ 656.003809] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 656.003809] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] raise exception.PortBindingFailed(port_id=port['id']) [ 656.004167] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] nova.exception.PortBindingFailed: Binding failed for port bfee4990-260b-40da-bbcf-869c85e6b40f, please check neutron logs for more information. [ 656.004167] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] [ 656.004167] env[63088]: INFO nova.compute.manager [None req-224c474a-3c21-4a3f-9526-5bb4e3ba7d78 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] Terminating instance [ 656.008478] env[63088]: DEBUG oslo_concurrency.lockutils [None req-224c474a-3c21-4a3f-9526-5bb4e3ba7d78 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] Acquiring lock "refresh_cache-62bd4013-f007-4f69-b037-14bedf1e63cf" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 656.100591] env[63088]: DEBUG nova.compute.utils [None req-cab9e1ae-5fd3-4816-8684-8d11c94f36d2 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 656.105341] env[63088]: DEBUG nova.compute.manager [None req-cab9e1ae-5fd3-4816-8684-8d11c94f36d2 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 656.105519] env[63088]: DEBUG nova.network.neutron [None req-cab9e1ae-5fd3-4816-8684-8d11c94f36d2 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 656.177124] env[63088]: DEBUG nova.policy [None req-cab9e1ae-5fd3-4816-8684-8d11c94f36d2 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '62988efd8a0f4819a2033dfac58f4670', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cf9a5b7b1a684d05afae04419343bd7e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 656.533609] env[63088]: DEBUG nova.network.neutron [req-250e3fc5-7138-4664-9bc2-9ce905f618b8 req-d2ecd7dd-ecad-4143-99aa-10d4aa97215c service nova] [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 656.606583] env[63088]: DEBUG nova.compute.manager [None req-cab9e1ae-5fd3-4816-8684-8d11c94f36d2 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 656.624341] env[63088]: DEBUG nova.network.neutron [None req-cab9e1ae-5fd3-4816-8684-8d11c94f36d2 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] Successfully created port: eb9c1776-bb3b-4b18-9c46-5dc6e7f78df5 {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 656.666635] env[63088]: DEBUG nova.network.neutron [req-250e3fc5-7138-4664-9bc2-9ce905f618b8 req-d2ecd7dd-ecad-4143-99aa-10d4aa97215c service nova] [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 657.043436] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5762c5a7-058d-45b4-bd01-d933f630d6e0 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.053904] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec753023-4dd0-4934-aaea-5c92cdbab33c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.093634] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2f4e1c1-f9f5-433a-afec-9fb523a96192 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.101264] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78be694c-8191-4322-97d8-82fe676baa9f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.114979] env[63088]: DEBUG nova.compute.provider_tree [None req-b7ea81d3-fe92-4160-8c6b-de9afb157d35 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 657.169036] env[63088]: DEBUG oslo_concurrency.lockutils [req-250e3fc5-7138-4664-9bc2-9ce905f618b8 req-d2ecd7dd-ecad-4143-99aa-10d4aa97215c service nova] Releasing lock "refresh_cache-62bd4013-f007-4f69-b037-14bedf1e63cf" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 657.169433] env[63088]: DEBUG oslo_concurrency.lockutils [None req-224c474a-3c21-4a3f-9526-5bb4e3ba7d78 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] Acquired lock "refresh_cache-62bd4013-f007-4f69-b037-14bedf1e63cf" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 657.169678] env[63088]: DEBUG nova.network.neutron [None req-224c474a-3c21-4a3f-9526-5bb4e3ba7d78 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 657.345199] env[63088]: DEBUG oslo_concurrency.lockutils [None req-1c0c7b6c-8782-4a52-b8ee-b7fd4463c345 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Acquiring lock "fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 657.345432] env[63088]: DEBUG oslo_concurrency.lockutils [None req-1c0c7b6c-8782-4a52-b8ee-b7fd4463c345 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Lock "fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 657.621709] env[63088]: DEBUG nova.scheduler.client.report [None req-b7ea81d3-fe92-4160-8c6b-de9afb157d35 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 657.630502] env[63088]: DEBUG nova.compute.manager [None req-cab9e1ae-5fd3-4816-8684-8d11c94f36d2 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 657.665287] env[63088]: DEBUG nova.virt.hardware [None req-cab9e1ae-5fd3-4816-8684-8d11c94f36d2 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 657.665519] env[63088]: DEBUG nova.virt.hardware [None req-cab9e1ae-5fd3-4816-8684-8d11c94f36d2 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 657.665880] env[63088]: DEBUG nova.virt.hardware [None req-cab9e1ae-5fd3-4816-8684-8d11c94f36d2 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 657.665880] env[63088]: DEBUG nova.virt.hardware [None req-cab9e1ae-5fd3-4816-8684-8d11c94f36d2 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 657.666134] env[63088]: DEBUG nova.virt.hardware [None req-cab9e1ae-5fd3-4816-8684-8d11c94f36d2 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 657.668328] env[63088]: DEBUG nova.virt.hardware [None req-cab9e1ae-5fd3-4816-8684-8d11c94f36d2 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 657.670345] env[63088]: DEBUG nova.virt.hardware [None req-cab9e1ae-5fd3-4816-8684-8d11c94f36d2 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 657.671058] env[63088]: DEBUG nova.virt.hardware [None req-cab9e1ae-5fd3-4816-8684-8d11c94f36d2 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 657.671058] env[63088]: DEBUG nova.virt.hardware [None req-cab9e1ae-5fd3-4816-8684-8d11c94f36d2 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 657.671058] env[63088]: DEBUG nova.virt.hardware [None req-cab9e1ae-5fd3-4816-8684-8d11c94f36d2 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 657.671154] env[63088]: DEBUG nova.virt.hardware [None req-cab9e1ae-5fd3-4816-8684-8d11c94f36d2 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 657.672449] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a7f3f3d-b4c9-436b-8e4a-5194ac5afa1e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.683852] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a5ee38b-1062-472e-85b6-ce986b2e6fb7 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.706540] env[63088]: DEBUG nova.network.neutron [None req-224c474a-3c21-4a3f-9526-5bb4e3ba7d78 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 657.793508] env[63088]: DEBUG nova.network.neutron [None req-224c474a-3c21-4a3f-9526-5bb4e3ba7d78 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 657.888255] env[63088]: DEBUG nova.compute.manager [req-415c3217-2eae-4069-8ca1-6a8024079017 req-289e0f86-d281-487f-af9c-f27d1eb31c30 service nova] [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] Received event network-vif-deleted-bfee4990-260b-40da-bbcf-869c85e6b40f {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 658.136014] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b7ea81d3-fe92-4160-8c6b-de9afb157d35 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.542s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 658.136638] env[63088]: DEBUG nova.compute.manager [None req-b7ea81d3-fe92-4160-8c6b-de9afb157d35 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 658.143022] env[63088]: DEBUG oslo_concurrency.lockutils [None req-442e0a41-8be2-46c5-9922-69b38e5ecc58 tempest-ServerAddressesNegativeTestJSON-62458283 tempest-ServerAddressesNegativeTestJSON-62458283-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.399s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 658.143022] env[63088]: INFO nova.compute.claims [None req-442e0a41-8be2-46c5-9922-69b38e5ecc58 tempest-ServerAddressesNegativeTestJSON-62458283 tempest-ServerAddressesNegativeTestJSON-62458283-project-member] [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 658.300980] env[63088]: DEBUG oslo_concurrency.lockutils [None req-224c474a-3c21-4a3f-9526-5bb4e3ba7d78 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] Releasing lock "refresh_cache-62bd4013-f007-4f69-b037-14bedf1e63cf" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 658.300980] env[63088]: DEBUG nova.compute.manager [None req-224c474a-3c21-4a3f-9526-5bb4e3ba7d78 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 658.300980] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-224c474a-3c21-4a3f-9526-5bb4e3ba7d78 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 658.300980] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7df2102d-3ea5-4df3-9dd3-e6af5c2f1a36 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.311633] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f33506ba-2182-48c1-bf4c-ae019db1a74f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.344508] env[63088]: WARNING nova.virt.vmwareapi.vmops [None req-224c474a-3c21-4a3f-9526-5bb4e3ba7d78 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 62bd4013-f007-4f69-b037-14bedf1e63cf could not be found. [ 658.345036] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-224c474a-3c21-4a3f-9526-5bb4e3ba7d78 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 658.345146] env[63088]: INFO nova.compute.manager [None req-224c474a-3c21-4a3f-9526-5bb4e3ba7d78 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] Took 0.05 seconds to destroy the instance on the hypervisor. [ 658.345535] env[63088]: DEBUG oslo.service.loopingcall [None req-224c474a-3c21-4a3f-9526-5bb4e3ba7d78 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 658.345945] env[63088]: DEBUG nova.compute.manager [-] [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 658.346093] env[63088]: DEBUG nova.network.neutron [-] [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 658.379963] env[63088]: DEBUG nova.network.neutron [-] [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 658.457658] env[63088]: ERROR nova.compute.manager [None req-cab9e1ae-5fd3-4816-8684-8d11c94f36d2 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port eb9c1776-bb3b-4b18-9c46-5dc6e7f78df5, please check neutron logs for more information. [ 658.457658] env[63088]: ERROR nova.compute.manager Traceback (most recent call last): [ 658.457658] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 658.457658] env[63088]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 658.457658] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 658.457658] env[63088]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 658.457658] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 658.457658] env[63088]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 658.457658] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 658.457658] env[63088]: ERROR nova.compute.manager self.force_reraise() [ 658.457658] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 658.457658] env[63088]: ERROR nova.compute.manager raise self.value [ 658.457658] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 658.457658] env[63088]: ERROR nova.compute.manager updated_port = self._update_port( [ 658.457658] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 658.457658] env[63088]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 658.458108] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 658.458108] env[63088]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 658.458108] env[63088]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port eb9c1776-bb3b-4b18-9c46-5dc6e7f78df5, please check neutron logs for more information. [ 658.458108] env[63088]: ERROR nova.compute.manager [ 658.458108] env[63088]: Traceback (most recent call last): [ 658.458108] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 658.458108] env[63088]: listener.cb(fileno) [ 658.458108] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 658.458108] env[63088]: result = function(*args, **kwargs) [ 658.458108] env[63088]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 658.458108] env[63088]: return func(*args, **kwargs) [ 658.458108] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 658.458108] env[63088]: raise e [ 658.458108] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 658.458108] env[63088]: nwinfo = self.network_api.allocate_for_instance( [ 658.458108] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 658.458108] env[63088]: created_port_ids = self._update_ports_for_instance( [ 658.458108] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 658.458108] env[63088]: with excutils.save_and_reraise_exception(): [ 658.458108] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 658.458108] env[63088]: self.force_reraise() [ 658.458108] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 658.458108] env[63088]: raise self.value [ 658.458108] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 658.458108] env[63088]: updated_port = self._update_port( [ 658.458108] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 658.458108] env[63088]: _ensure_no_port_binding_failure(port) [ 658.458108] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 658.458108] env[63088]: raise exception.PortBindingFailed(port_id=port['id']) [ 658.458966] env[63088]: nova.exception.PortBindingFailed: Binding failed for port eb9c1776-bb3b-4b18-9c46-5dc6e7f78df5, please check neutron logs for more information. [ 658.458966] env[63088]: Removing descriptor: 15 [ 658.460539] env[63088]: ERROR nova.compute.manager [None req-cab9e1ae-5fd3-4816-8684-8d11c94f36d2 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port eb9c1776-bb3b-4b18-9c46-5dc6e7f78df5, please check neutron logs for more information. [ 658.460539] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] Traceback (most recent call last): [ 658.460539] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 658.460539] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] yield resources [ 658.460539] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 658.460539] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] self.driver.spawn(context, instance, image_meta, [ 658.460539] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 658.460539] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] self._vmops.spawn(context, instance, image_meta, injected_files, [ 658.460539] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 658.460539] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] vm_ref = self.build_virtual_machine(instance, [ 658.460539] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 658.460920] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] vif_infos = vmwarevif.get_vif_info(self._session, [ 658.460920] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 658.460920] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] for vif in network_info: [ 658.460920] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 658.460920] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] return self._sync_wrapper(fn, *args, **kwargs) [ 658.460920] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 658.460920] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] self.wait() [ 658.460920] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 658.460920] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] self[:] = self._gt.wait() [ 658.460920] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 658.460920] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] return self._exit_event.wait() [ 658.460920] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 658.460920] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] result = hub.switch() [ 658.461283] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 658.461283] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] return self.greenlet.switch() [ 658.461283] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 658.461283] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] result = function(*args, **kwargs) [ 658.461283] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 658.461283] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] return func(*args, **kwargs) [ 658.461283] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 658.461283] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] raise e [ 658.461283] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 658.461283] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] nwinfo = self.network_api.allocate_for_instance( [ 658.461283] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 658.461283] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] created_port_ids = self._update_ports_for_instance( [ 658.461283] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 658.461640] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] with excutils.save_and_reraise_exception(): [ 658.461640] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 658.461640] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] self.force_reraise() [ 658.461640] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 658.461640] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] raise self.value [ 658.461640] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 658.461640] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] updated_port = self._update_port( [ 658.461640] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 658.461640] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] _ensure_no_port_binding_failure(port) [ 658.461640] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 658.461640] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] raise exception.PortBindingFailed(port_id=port['id']) [ 658.461640] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] nova.exception.PortBindingFailed: Binding failed for port eb9c1776-bb3b-4b18-9c46-5dc6e7f78df5, please check neutron logs for more information. [ 658.461640] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] [ 658.462023] env[63088]: INFO nova.compute.manager [None req-cab9e1ae-5fd3-4816-8684-8d11c94f36d2 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] Terminating instance [ 658.468167] env[63088]: DEBUG oslo_concurrency.lockutils [None req-cab9e1ae-5fd3-4816-8684-8d11c94f36d2 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] Acquiring lock "refresh_cache-a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 658.468336] env[63088]: DEBUG oslo_concurrency.lockutils [None req-cab9e1ae-5fd3-4816-8684-8d11c94f36d2 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] Acquired lock "refresh_cache-a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 658.468512] env[63088]: DEBUG nova.network.neutron [None req-cab9e1ae-5fd3-4816-8684-8d11c94f36d2 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 658.647205] env[63088]: DEBUG nova.compute.utils [None req-b7ea81d3-fe92-4160-8c6b-de9afb157d35 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 658.648794] env[63088]: DEBUG nova.compute.manager [None req-b7ea81d3-fe92-4160-8c6b-de9afb157d35 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 658.649082] env[63088]: DEBUG nova.network.neutron [None req-b7ea81d3-fe92-4160-8c6b-de9afb157d35 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 658.708029] env[63088]: DEBUG nova.policy [None req-b7ea81d3-fe92-4160-8c6b-de9afb157d35 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '62988efd8a0f4819a2033dfac58f4670', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cf9a5b7b1a684d05afae04419343bd7e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 658.883700] env[63088]: DEBUG nova.network.neutron [-] [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 659.018148] env[63088]: DEBUG nova.network.neutron [None req-cab9e1ae-5fd3-4816-8684-8d11c94f36d2 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 659.054500] env[63088]: DEBUG nova.network.neutron [None req-b7ea81d3-fe92-4160-8c6b-de9afb157d35 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] Successfully created port: 7c5a6f5a-771f-481d-a6f2-f56e1d85dfc9 {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 659.092554] env[63088]: DEBUG nova.network.neutron [None req-cab9e1ae-5fd3-4816-8684-8d11c94f36d2 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 659.149683] env[63088]: DEBUG nova.compute.manager [None req-b7ea81d3-fe92-4160-8c6b-de9afb157d35 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 659.388598] env[63088]: INFO nova.compute.manager [-] [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] Took 1.04 seconds to deallocate network for instance. [ 659.391974] env[63088]: DEBUG nova.compute.claims [None req-224c474a-3c21-4a3f-9526-5bb4e3ba7d78 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] Aborting claim: {{(pid=63088) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 659.391974] env[63088]: DEBUG oslo_concurrency.lockutils [None req-224c474a-3c21-4a3f-9526-5bb4e3ba7d78 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 659.574297] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcd86c61-b864-4894-bb64-d919ecf4f8c1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.583392] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d92b204-d816-4159-9077-8f151de0711f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.615024] env[63088]: DEBUG oslo_concurrency.lockutils [None req-cab9e1ae-5fd3-4816-8684-8d11c94f36d2 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] Releasing lock "refresh_cache-a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 659.615290] env[63088]: DEBUG nova.compute.manager [None req-cab9e1ae-5fd3-4816-8684-8d11c94f36d2 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 659.615494] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-cab9e1ae-5fd3-4816-8684-8d11c94f36d2 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 659.615946] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9ebb7416-ebfe-4ea1-8a8b-3e056c091839 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.618289] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b020ac1b-9f4b-44bf-9393-014e9f536c69 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.627247] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08da9efa-ae3c-4e86-98dc-fa7d45ac7bcb {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.633599] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4041235-44b2-475d-8836-d101013f69f5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.658378] env[63088]: DEBUG nova.compute.provider_tree [None req-442e0a41-8be2-46c5-9922-69b38e5ecc58 tempest-ServerAddressesNegativeTestJSON-62458283 tempest-ServerAddressesNegativeTestJSON-62458283-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 659.664705] env[63088]: WARNING nova.virt.vmwareapi.vmops [None req-cab9e1ae-5fd3-4816-8684-8d11c94f36d2 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af could not be found. [ 659.664705] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-cab9e1ae-5fd3-4816-8684-8d11c94f36d2 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 659.664705] env[63088]: INFO nova.compute.manager [None req-cab9e1ae-5fd3-4816-8684-8d11c94f36d2 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] Took 0.05 seconds to destroy the instance on the hypervisor. [ 659.664705] env[63088]: DEBUG oslo.service.loopingcall [None req-cab9e1ae-5fd3-4816-8684-8d11c94f36d2 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 659.664705] env[63088]: DEBUG nova.compute.manager [-] [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 659.664705] env[63088]: DEBUG nova.network.neutron [-] [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 659.684248] env[63088]: DEBUG nova.network.neutron [-] [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 659.916619] env[63088]: DEBUG nova.compute.manager [req-0f8f6cc0-7e9b-4ed4-8f0f-03bc8812addd req-bb8f0c5f-2bbd-46c2-9693-3d6ee554f692 service nova] [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] Received event network-changed-eb9c1776-bb3b-4b18-9c46-5dc6e7f78df5 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 659.916821] env[63088]: DEBUG nova.compute.manager [req-0f8f6cc0-7e9b-4ed4-8f0f-03bc8812addd req-bb8f0c5f-2bbd-46c2-9693-3d6ee554f692 service nova] [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] Refreshing instance network info cache due to event network-changed-eb9c1776-bb3b-4b18-9c46-5dc6e7f78df5. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 659.917040] env[63088]: DEBUG oslo_concurrency.lockutils [req-0f8f6cc0-7e9b-4ed4-8f0f-03bc8812addd req-bb8f0c5f-2bbd-46c2-9693-3d6ee554f692 service nova] Acquiring lock "refresh_cache-a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 659.917186] env[63088]: DEBUG oslo_concurrency.lockutils [req-0f8f6cc0-7e9b-4ed4-8f0f-03bc8812addd req-bb8f0c5f-2bbd-46c2-9693-3d6ee554f692 service nova] Acquired lock "refresh_cache-a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 659.917346] env[63088]: DEBUG nova.network.neutron [req-0f8f6cc0-7e9b-4ed4-8f0f-03bc8812addd req-bb8f0c5f-2bbd-46c2-9693-3d6ee554f692 service nova] [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] Refreshing network info cache for port eb9c1776-bb3b-4b18-9c46-5dc6e7f78df5 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 660.165409] env[63088]: DEBUG nova.compute.manager [None req-b7ea81d3-fe92-4160-8c6b-de9afb157d35 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 660.168988] env[63088]: DEBUG nova.scheduler.client.report [None req-442e0a41-8be2-46c5-9922-69b38e5ecc58 tempest-ServerAddressesNegativeTestJSON-62458283 tempest-ServerAddressesNegativeTestJSON-62458283-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 660.183793] env[63088]: DEBUG nova.network.neutron [-] [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 660.196091] env[63088]: DEBUG nova.virt.hardware [None req-b7ea81d3-fe92-4160-8c6b-de9afb157d35 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 660.196091] env[63088]: DEBUG nova.virt.hardware [None req-b7ea81d3-fe92-4160-8c6b-de9afb157d35 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 660.196091] env[63088]: DEBUG nova.virt.hardware [None req-b7ea81d3-fe92-4160-8c6b-de9afb157d35 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 660.196884] env[63088]: DEBUG nova.virt.hardware [None req-b7ea81d3-fe92-4160-8c6b-de9afb157d35 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 660.196956] env[63088]: DEBUG nova.virt.hardware [None req-b7ea81d3-fe92-4160-8c6b-de9afb157d35 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 660.197284] env[63088]: DEBUG nova.virt.hardware [None req-b7ea81d3-fe92-4160-8c6b-de9afb157d35 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 660.197460] env[63088]: DEBUG nova.virt.hardware [None req-b7ea81d3-fe92-4160-8c6b-de9afb157d35 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 660.197709] env[63088]: DEBUG nova.virt.hardware [None req-b7ea81d3-fe92-4160-8c6b-de9afb157d35 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 660.198515] env[63088]: DEBUG nova.virt.hardware [None req-b7ea81d3-fe92-4160-8c6b-de9afb157d35 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 660.198515] env[63088]: DEBUG nova.virt.hardware [None req-b7ea81d3-fe92-4160-8c6b-de9afb157d35 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 660.198515] env[63088]: DEBUG nova.virt.hardware [None req-b7ea81d3-fe92-4160-8c6b-de9afb157d35 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 660.202226] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-053aedee-1fc6-4968-9776-a7628e02defd {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.209510] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5afb2b22-ce9c-4051-83a5-719963b8ee1c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.238761] env[63088]: ERROR nova.compute.manager [None req-b7ea81d3-fe92-4160-8c6b-de9afb157d35 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 7c5a6f5a-771f-481d-a6f2-f56e1d85dfc9, please check neutron logs for more information. [ 660.238761] env[63088]: ERROR nova.compute.manager Traceback (most recent call last): [ 660.238761] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 660.238761] env[63088]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 660.238761] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 660.238761] env[63088]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 660.238761] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 660.238761] env[63088]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 660.238761] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 660.238761] env[63088]: ERROR nova.compute.manager self.force_reraise() [ 660.238761] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 660.238761] env[63088]: ERROR nova.compute.manager raise self.value [ 660.238761] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 660.238761] env[63088]: ERROR nova.compute.manager updated_port = self._update_port( [ 660.238761] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 660.238761] env[63088]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 660.239240] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 660.239240] env[63088]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 660.239240] env[63088]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 7c5a6f5a-771f-481d-a6f2-f56e1d85dfc9, please check neutron logs for more information. [ 660.239240] env[63088]: ERROR nova.compute.manager [ 660.239240] env[63088]: Traceback (most recent call last): [ 660.239240] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 660.239240] env[63088]: listener.cb(fileno) [ 660.239240] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 660.239240] env[63088]: result = function(*args, **kwargs) [ 660.239240] env[63088]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 660.239240] env[63088]: return func(*args, **kwargs) [ 660.239240] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 660.239240] env[63088]: raise e [ 660.239240] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 660.239240] env[63088]: nwinfo = self.network_api.allocate_for_instance( [ 660.239240] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 660.239240] env[63088]: created_port_ids = self._update_ports_for_instance( [ 660.239240] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 660.239240] env[63088]: with excutils.save_and_reraise_exception(): [ 660.239240] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 660.239240] env[63088]: self.force_reraise() [ 660.239240] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 660.239240] env[63088]: raise self.value [ 660.239240] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 660.239240] env[63088]: updated_port = self._update_port( [ 660.239240] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 660.239240] env[63088]: _ensure_no_port_binding_failure(port) [ 660.239240] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 660.239240] env[63088]: raise exception.PortBindingFailed(port_id=port['id']) [ 660.239918] env[63088]: nova.exception.PortBindingFailed: Binding failed for port 7c5a6f5a-771f-481d-a6f2-f56e1d85dfc9, please check neutron logs for more information. [ 660.239918] env[63088]: Removing descriptor: 15 [ 660.239918] env[63088]: ERROR nova.compute.manager [None req-b7ea81d3-fe92-4160-8c6b-de9afb157d35 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 7c5a6f5a-771f-481d-a6f2-f56e1d85dfc9, please check neutron logs for more information. [ 660.239918] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] Traceback (most recent call last): [ 660.239918] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 660.239918] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] yield resources [ 660.239918] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 660.239918] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] self.driver.spawn(context, instance, image_meta, [ 660.239918] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 660.239918] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] self._vmops.spawn(context, instance, image_meta, injected_files, [ 660.239918] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 660.239918] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] vm_ref = self.build_virtual_machine(instance, [ 660.241049] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 660.241049] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] vif_infos = vmwarevif.get_vif_info(self._session, [ 660.241049] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 660.241049] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] for vif in network_info: [ 660.241049] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 660.241049] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] return self._sync_wrapper(fn, *args, **kwargs) [ 660.241049] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 660.241049] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] self.wait() [ 660.241049] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 660.241049] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] self[:] = self._gt.wait() [ 660.241049] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 660.241049] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] return self._exit_event.wait() [ 660.241049] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 660.241346] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] result = hub.switch() [ 660.241346] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 660.241346] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] return self.greenlet.switch() [ 660.241346] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 660.241346] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] result = function(*args, **kwargs) [ 660.241346] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 660.241346] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] return func(*args, **kwargs) [ 660.241346] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 660.241346] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] raise e [ 660.241346] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 660.241346] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] nwinfo = self.network_api.allocate_for_instance( [ 660.241346] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 660.241346] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] created_port_ids = self._update_ports_for_instance( [ 660.243249] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 660.243249] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] with excutils.save_and_reraise_exception(): [ 660.243249] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 660.243249] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] self.force_reraise() [ 660.243249] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 660.243249] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] raise self.value [ 660.243249] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 660.243249] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] updated_port = self._update_port( [ 660.243249] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 660.243249] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] _ensure_no_port_binding_failure(port) [ 660.243249] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 660.243249] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] raise exception.PortBindingFailed(port_id=port['id']) [ 660.243981] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] nova.exception.PortBindingFailed: Binding failed for port 7c5a6f5a-771f-481d-a6f2-f56e1d85dfc9, please check neutron logs for more information. [ 660.243981] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] [ 660.243981] env[63088]: INFO nova.compute.manager [None req-b7ea81d3-fe92-4160-8c6b-de9afb157d35 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] Terminating instance [ 660.243981] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b7ea81d3-fe92-4160-8c6b-de9afb157d35 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] Acquiring lock "refresh_cache-945b5a77-96b9-4565-94a8-6f1a89dc3bbf" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 660.243981] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b7ea81d3-fe92-4160-8c6b-de9afb157d35 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] Acquired lock "refresh_cache-945b5a77-96b9-4565-94a8-6f1a89dc3bbf" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 660.243981] env[63088]: DEBUG nova.network.neutron [None req-b7ea81d3-fe92-4160-8c6b-de9afb157d35 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 660.263952] env[63088]: DEBUG oslo_concurrency.lockutils [None req-96c00a75-0383-4506-a4c6-d540aa18aaef tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Acquiring lock "87361d1c-e9cd-4177-bf3d-5d7551192073" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 660.264164] env[63088]: DEBUG oslo_concurrency.lockutils [None req-96c00a75-0383-4506-a4c6-d540aa18aaef tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Lock "87361d1c-e9cd-4177-bf3d-5d7551192073" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 660.435331] env[63088]: DEBUG nova.network.neutron [req-0f8f6cc0-7e9b-4ed4-8f0f-03bc8812addd req-bb8f0c5f-2bbd-46c2-9693-3d6ee554f692 service nova] [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 660.484843] env[63088]: DEBUG nova.network.neutron [req-0f8f6cc0-7e9b-4ed4-8f0f-03bc8812addd req-bb8f0c5f-2bbd-46c2-9693-3d6ee554f692 service nova] [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 660.678051] env[63088]: DEBUG oslo_concurrency.lockutils [None req-442e0a41-8be2-46c5-9922-69b38e5ecc58 tempest-ServerAddressesNegativeTestJSON-62458283 tempest-ServerAddressesNegativeTestJSON-62458283-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.535s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 660.678051] env[63088]: DEBUG nova.compute.manager [None req-442e0a41-8be2-46c5-9922-69b38e5ecc58 tempest-ServerAddressesNegativeTestJSON-62458283 tempest-ServerAddressesNegativeTestJSON-62458283-project-member] [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 660.678449] env[63088]: DEBUG oslo_concurrency.lockutils [None req-32cec7a8-18a7-43df-b82f-d466be0dcd4a tempest-AttachInterfacesUnderV243Test-1205880501 tempest-AttachInterfacesUnderV243Test-1205880501-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.601s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 660.679929] env[63088]: INFO nova.compute.claims [None req-32cec7a8-18a7-43df-b82f-d466be0dcd4a tempest-AttachInterfacesUnderV243Test-1205880501 tempest-AttachInterfacesUnderV243Test-1205880501-project-member] [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 660.686029] env[63088]: INFO nova.compute.manager [-] [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] Took 1.02 seconds to deallocate network for instance. [ 660.688531] env[63088]: DEBUG nova.compute.claims [None req-cab9e1ae-5fd3-4816-8684-8d11c94f36d2 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] Aborting claim: {{(pid=63088) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 660.688770] env[63088]: DEBUG oslo_concurrency.lockutils [None req-cab9e1ae-5fd3-4816-8684-8d11c94f36d2 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 660.759832] env[63088]: DEBUG nova.network.neutron [None req-b7ea81d3-fe92-4160-8c6b-de9afb157d35 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 660.830872] env[63088]: DEBUG nova.network.neutron [None req-b7ea81d3-fe92-4160-8c6b-de9afb157d35 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 660.987770] env[63088]: DEBUG oslo_concurrency.lockutils [req-0f8f6cc0-7e9b-4ed4-8f0f-03bc8812addd req-bb8f0c5f-2bbd-46c2-9693-3d6ee554f692 service nova] Releasing lock "refresh_cache-a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 660.988064] env[63088]: DEBUG nova.compute.manager [req-0f8f6cc0-7e9b-4ed4-8f0f-03bc8812addd req-bb8f0c5f-2bbd-46c2-9693-3d6ee554f692 service nova] [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] Received event network-vif-deleted-eb9c1776-bb3b-4b18-9c46-5dc6e7f78df5 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 661.184440] env[63088]: DEBUG nova.compute.utils [None req-442e0a41-8be2-46c5-9922-69b38e5ecc58 tempest-ServerAddressesNegativeTestJSON-62458283 tempest-ServerAddressesNegativeTestJSON-62458283-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 661.187737] env[63088]: DEBUG nova.compute.manager [None req-442e0a41-8be2-46c5-9922-69b38e5ecc58 tempest-ServerAddressesNegativeTestJSON-62458283 tempest-ServerAddressesNegativeTestJSON-62458283-project-member] [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 661.187906] env[63088]: DEBUG nova.network.neutron [None req-442e0a41-8be2-46c5-9922-69b38e5ecc58 tempest-ServerAddressesNegativeTestJSON-62458283 tempest-ServerAddressesNegativeTestJSON-62458283-project-member] [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 661.307776] env[63088]: DEBUG nova.policy [None req-442e0a41-8be2-46c5-9922-69b38e5ecc58 tempest-ServerAddressesNegativeTestJSON-62458283 tempest-ServerAddressesNegativeTestJSON-62458283-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '12e8946617cc438b854e0d94bcdab894', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '937f7a2fd5424bac8fd497f89b435c13', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 661.333556] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b7ea81d3-fe92-4160-8c6b-de9afb157d35 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] Releasing lock "refresh_cache-945b5a77-96b9-4565-94a8-6f1a89dc3bbf" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 661.334077] env[63088]: DEBUG nova.compute.manager [None req-b7ea81d3-fe92-4160-8c6b-de9afb157d35 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 661.334278] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-b7ea81d3-fe92-4160-8c6b-de9afb157d35 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 661.334567] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c9801dec-7738-4e9c-9184-8df92e3a2ead {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.344508] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e276bd7-2c77-452a-96c6-88566be22ca6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.368844] env[63088]: WARNING nova.virt.vmwareapi.vmops [None req-b7ea81d3-fe92-4160-8c6b-de9afb157d35 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 945b5a77-96b9-4565-94a8-6f1a89dc3bbf could not be found. [ 661.369157] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-b7ea81d3-fe92-4160-8c6b-de9afb157d35 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 661.369358] env[63088]: INFO nova.compute.manager [None req-b7ea81d3-fe92-4160-8c6b-de9afb157d35 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] Took 0.04 seconds to destroy the instance on the hypervisor. [ 661.369618] env[63088]: DEBUG oslo.service.loopingcall [None req-b7ea81d3-fe92-4160-8c6b-de9afb157d35 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 661.369882] env[63088]: DEBUG nova.compute.manager [-] [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 661.369975] env[63088]: DEBUG nova.network.neutron [-] [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 661.383931] env[63088]: DEBUG nova.network.neutron [-] [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 661.572891] env[63088]: DEBUG nova.network.neutron [None req-442e0a41-8be2-46c5-9922-69b38e5ecc58 tempest-ServerAddressesNegativeTestJSON-62458283 tempest-ServerAddressesNegativeTestJSON-62458283-project-member] [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] Successfully created port: be2d3e68-49fa-4eae-8386-e494fa78b11b {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 661.693728] env[63088]: DEBUG nova.compute.manager [None req-442e0a41-8be2-46c5-9922-69b38e5ecc58 tempest-ServerAddressesNegativeTestJSON-62458283 tempest-ServerAddressesNegativeTestJSON-62458283-project-member] [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 661.889031] env[63088]: DEBUG nova.network.neutron [-] [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 661.979011] env[63088]: DEBUG nova.compute.manager [req-473d62aa-b91f-471b-8f36-c662ff057791 req-fa75a089-f00d-42cb-867b-b014d2b25073 service nova] [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] Received event network-changed-7c5a6f5a-771f-481d-a6f2-f56e1d85dfc9 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 661.979011] env[63088]: DEBUG nova.compute.manager [req-473d62aa-b91f-471b-8f36-c662ff057791 req-fa75a089-f00d-42cb-867b-b014d2b25073 service nova] [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] Refreshing instance network info cache due to event network-changed-7c5a6f5a-771f-481d-a6f2-f56e1d85dfc9. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 661.979011] env[63088]: DEBUG oslo_concurrency.lockutils [req-473d62aa-b91f-471b-8f36-c662ff057791 req-fa75a089-f00d-42cb-867b-b014d2b25073 service nova] Acquiring lock "refresh_cache-945b5a77-96b9-4565-94a8-6f1a89dc3bbf" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 661.979011] env[63088]: DEBUG oslo_concurrency.lockutils [req-473d62aa-b91f-471b-8f36-c662ff057791 req-fa75a089-f00d-42cb-867b-b014d2b25073 service nova] Acquired lock "refresh_cache-945b5a77-96b9-4565-94a8-6f1a89dc3bbf" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 661.979470] env[63088]: DEBUG nova.network.neutron [req-473d62aa-b91f-471b-8f36-c662ff057791 req-fa75a089-f00d-42cb-867b-b014d2b25073 service nova] [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] Refreshing network info cache for port 7c5a6f5a-771f-481d-a6f2-f56e1d85dfc9 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 662.110904] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95caa49a-7e81-44d5-b69c-204995bf9d36 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.123018] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2d25dd4-cf06-4ea0-9026-78b21610fa31 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.150342] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9089793b-6417-4a25-b460-41b117d8231d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.158354] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bac5aa8-25e9-4c10-8827-70c3ddca317d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.171882] env[63088]: DEBUG nova.compute.provider_tree [None req-32cec7a8-18a7-43df-b82f-d466be0dcd4a tempest-AttachInterfacesUnderV243Test-1205880501 tempest-AttachInterfacesUnderV243Test-1205880501-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 662.397651] env[63088]: INFO nova.compute.manager [-] [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] Took 1.03 seconds to deallocate network for instance. [ 662.400121] env[63088]: DEBUG nova.compute.claims [None req-b7ea81d3-fe92-4160-8c6b-de9afb157d35 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] Aborting claim: {{(pid=63088) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 662.400326] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b7ea81d3-fe92-4160-8c6b-de9afb157d35 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 662.499395] env[63088]: DEBUG nova.network.neutron [req-473d62aa-b91f-471b-8f36-c662ff057791 req-fa75a089-f00d-42cb-867b-b014d2b25073 service nova] [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 662.520884] env[63088]: ERROR nova.compute.manager [None req-442e0a41-8be2-46c5-9922-69b38e5ecc58 tempest-ServerAddressesNegativeTestJSON-62458283 tempest-ServerAddressesNegativeTestJSON-62458283-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port be2d3e68-49fa-4eae-8386-e494fa78b11b, please check neutron logs for more information. [ 662.520884] env[63088]: ERROR nova.compute.manager Traceback (most recent call last): [ 662.520884] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 662.520884] env[63088]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 662.520884] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 662.520884] env[63088]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 662.520884] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 662.520884] env[63088]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 662.520884] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 662.520884] env[63088]: ERROR nova.compute.manager self.force_reraise() [ 662.520884] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 662.520884] env[63088]: ERROR nova.compute.manager raise self.value [ 662.520884] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 662.520884] env[63088]: ERROR nova.compute.manager updated_port = self._update_port( [ 662.520884] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 662.520884] env[63088]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 662.521355] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 662.521355] env[63088]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 662.521355] env[63088]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port be2d3e68-49fa-4eae-8386-e494fa78b11b, please check neutron logs for more information. [ 662.521355] env[63088]: ERROR nova.compute.manager [ 662.521355] env[63088]: Traceback (most recent call last): [ 662.521355] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 662.521355] env[63088]: listener.cb(fileno) [ 662.521355] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 662.521355] env[63088]: result = function(*args, **kwargs) [ 662.521355] env[63088]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 662.521355] env[63088]: return func(*args, **kwargs) [ 662.521355] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 662.521355] env[63088]: raise e [ 662.521355] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 662.521355] env[63088]: nwinfo = self.network_api.allocate_for_instance( [ 662.521355] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 662.521355] env[63088]: created_port_ids = self._update_ports_for_instance( [ 662.521355] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 662.521355] env[63088]: with excutils.save_and_reraise_exception(): [ 662.521355] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 662.521355] env[63088]: self.force_reraise() [ 662.521355] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 662.521355] env[63088]: raise self.value [ 662.521355] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 662.521355] env[63088]: updated_port = self._update_port( [ 662.521355] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 662.521355] env[63088]: _ensure_no_port_binding_failure(port) [ 662.521355] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 662.521355] env[63088]: raise exception.PortBindingFailed(port_id=port['id']) [ 662.522240] env[63088]: nova.exception.PortBindingFailed: Binding failed for port be2d3e68-49fa-4eae-8386-e494fa78b11b, please check neutron logs for more information. [ 662.522240] env[63088]: Removing descriptor: 15 [ 662.553093] env[63088]: DEBUG nova.network.neutron [req-473d62aa-b91f-471b-8f36-c662ff057791 req-fa75a089-f00d-42cb-867b-b014d2b25073 service nova] [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 662.675183] env[63088]: DEBUG nova.scheduler.client.report [None req-32cec7a8-18a7-43df-b82f-d466be0dcd4a tempest-AttachInterfacesUnderV243Test-1205880501 tempest-AttachInterfacesUnderV243Test-1205880501-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 662.706335] env[63088]: DEBUG nova.compute.manager [None req-442e0a41-8be2-46c5-9922-69b38e5ecc58 tempest-ServerAddressesNegativeTestJSON-62458283 tempest-ServerAddressesNegativeTestJSON-62458283-project-member] [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 662.732216] env[63088]: DEBUG nova.virt.hardware [None req-442e0a41-8be2-46c5-9922-69b38e5ecc58 tempest-ServerAddressesNegativeTestJSON-62458283 tempest-ServerAddressesNegativeTestJSON-62458283-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 662.732461] env[63088]: DEBUG nova.virt.hardware [None req-442e0a41-8be2-46c5-9922-69b38e5ecc58 tempest-ServerAddressesNegativeTestJSON-62458283 tempest-ServerAddressesNegativeTestJSON-62458283-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 662.732618] env[63088]: DEBUG nova.virt.hardware [None req-442e0a41-8be2-46c5-9922-69b38e5ecc58 tempest-ServerAddressesNegativeTestJSON-62458283 tempest-ServerAddressesNegativeTestJSON-62458283-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 662.732799] env[63088]: DEBUG nova.virt.hardware [None req-442e0a41-8be2-46c5-9922-69b38e5ecc58 tempest-ServerAddressesNegativeTestJSON-62458283 tempest-ServerAddressesNegativeTestJSON-62458283-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 662.732942] env[63088]: DEBUG nova.virt.hardware [None req-442e0a41-8be2-46c5-9922-69b38e5ecc58 tempest-ServerAddressesNegativeTestJSON-62458283 tempest-ServerAddressesNegativeTestJSON-62458283-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 662.733099] env[63088]: DEBUG nova.virt.hardware [None req-442e0a41-8be2-46c5-9922-69b38e5ecc58 tempest-ServerAddressesNegativeTestJSON-62458283 tempest-ServerAddressesNegativeTestJSON-62458283-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 662.733314] env[63088]: DEBUG nova.virt.hardware [None req-442e0a41-8be2-46c5-9922-69b38e5ecc58 tempest-ServerAddressesNegativeTestJSON-62458283 tempest-ServerAddressesNegativeTestJSON-62458283-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 662.733463] env[63088]: DEBUG nova.virt.hardware [None req-442e0a41-8be2-46c5-9922-69b38e5ecc58 tempest-ServerAddressesNegativeTestJSON-62458283 tempest-ServerAddressesNegativeTestJSON-62458283-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 662.733622] env[63088]: DEBUG nova.virt.hardware [None req-442e0a41-8be2-46c5-9922-69b38e5ecc58 tempest-ServerAddressesNegativeTestJSON-62458283 tempest-ServerAddressesNegativeTestJSON-62458283-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 662.733782] env[63088]: DEBUG nova.virt.hardware [None req-442e0a41-8be2-46c5-9922-69b38e5ecc58 tempest-ServerAddressesNegativeTestJSON-62458283 tempest-ServerAddressesNegativeTestJSON-62458283-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 662.733949] env[63088]: DEBUG nova.virt.hardware [None req-442e0a41-8be2-46c5-9922-69b38e5ecc58 tempest-ServerAddressesNegativeTestJSON-62458283 tempest-ServerAddressesNegativeTestJSON-62458283-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 662.734830] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-949c52d7-b0b1-442f-aff6-9d3263fd38e7 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.743101] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40ceae37-6ace-49ff-b43c-b66892c31837 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.760196] env[63088]: ERROR nova.compute.manager [None req-442e0a41-8be2-46c5-9922-69b38e5ecc58 tempest-ServerAddressesNegativeTestJSON-62458283 tempest-ServerAddressesNegativeTestJSON-62458283-project-member] [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port be2d3e68-49fa-4eae-8386-e494fa78b11b, please check neutron logs for more information. [ 662.760196] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] Traceback (most recent call last): [ 662.760196] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 662.760196] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] yield resources [ 662.760196] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 662.760196] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] self.driver.spawn(context, instance, image_meta, [ 662.760196] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 662.760196] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 662.760196] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 662.760196] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] vm_ref = self.build_virtual_machine(instance, [ 662.760196] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 662.760477] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] vif_infos = vmwarevif.get_vif_info(self._session, [ 662.760477] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 662.760477] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] for vif in network_info: [ 662.760477] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 662.760477] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] return self._sync_wrapper(fn, *args, **kwargs) [ 662.760477] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 662.760477] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] self.wait() [ 662.760477] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 662.760477] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] self[:] = self._gt.wait() [ 662.760477] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 662.760477] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] return self._exit_event.wait() [ 662.760477] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 662.760477] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] current.throw(*self._exc) [ 662.760976] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 662.760976] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] result = function(*args, **kwargs) [ 662.760976] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 662.760976] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] return func(*args, **kwargs) [ 662.760976] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 662.760976] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] raise e [ 662.760976] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 662.760976] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] nwinfo = self.network_api.allocate_for_instance( [ 662.760976] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 662.760976] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] created_port_ids = self._update_ports_for_instance( [ 662.760976] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 662.760976] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] with excutils.save_and_reraise_exception(): [ 662.760976] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 662.761386] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] self.force_reraise() [ 662.761386] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 662.761386] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] raise self.value [ 662.761386] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 662.761386] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] updated_port = self._update_port( [ 662.761386] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 662.761386] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] _ensure_no_port_binding_failure(port) [ 662.761386] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 662.761386] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] raise exception.PortBindingFailed(port_id=port['id']) [ 662.761386] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] nova.exception.PortBindingFailed: Binding failed for port be2d3e68-49fa-4eae-8386-e494fa78b11b, please check neutron logs for more information. [ 662.761386] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] [ 662.761386] env[63088]: INFO nova.compute.manager [None req-442e0a41-8be2-46c5-9922-69b38e5ecc58 tempest-ServerAddressesNegativeTestJSON-62458283 tempest-ServerAddressesNegativeTestJSON-62458283-project-member] [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] Terminating instance [ 662.762415] env[63088]: DEBUG oslo_concurrency.lockutils [None req-442e0a41-8be2-46c5-9922-69b38e5ecc58 tempest-ServerAddressesNegativeTestJSON-62458283 tempest-ServerAddressesNegativeTestJSON-62458283-project-member] Acquiring lock "refresh_cache-be5c7b15-c036-4a20-a21b-74bba67c9a2c" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 662.762604] env[63088]: DEBUG oslo_concurrency.lockutils [None req-442e0a41-8be2-46c5-9922-69b38e5ecc58 tempest-ServerAddressesNegativeTestJSON-62458283 tempest-ServerAddressesNegativeTestJSON-62458283-project-member] Acquired lock "refresh_cache-be5c7b15-c036-4a20-a21b-74bba67c9a2c" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 662.762727] env[63088]: DEBUG nova.network.neutron [None req-442e0a41-8be2-46c5-9922-69b38e5ecc58 tempest-ServerAddressesNegativeTestJSON-62458283 tempest-ServerAddressesNegativeTestJSON-62458283-project-member] [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 663.056706] env[63088]: DEBUG oslo_concurrency.lockutils [req-473d62aa-b91f-471b-8f36-c662ff057791 req-fa75a089-f00d-42cb-867b-b014d2b25073 service nova] Releasing lock "refresh_cache-945b5a77-96b9-4565-94a8-6f1a89dc3bbf" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 663.056978] env[63088]: DEBUG nova.compute.manager [req-473d62aa-b91f-471b-8f36-c662ff057791 req-fa75a089-f00d-42cb-867b-b014d2b25073 service nova] [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] Received event network-vif-deleted-7c5a6f5a-771f-481d-a6f2-f56e1d85dfc9 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 663.180331] env[63088]: DEBUG oslo_concurrency.lockutils [None req-32cec7a8-18a7-43df-b82f-d466be0dcd4a tempest-AttachInterfacesUnderV243Test-1205880501 tempest-AttachInterfacesUnderV243Test-1205880501-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.502s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 663.180880] env[63088]: DEBUG nova.compute.manager [None req-32cec7a8-18a7-43df-b82f-d466be0dcd4a tempest-AttachInterfacesUnderV243Test-1205880501 tempest-AttachInterfacesUnderV243Test-1205880501-project-member] [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 663.183560] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bb8843aa-5c19-4289-a41c-17cbfcb47a00 tempest-ServerActionsTestOtherA-1614359098 tempest-ServerActionsTestOtherA-1614359098-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.455s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 663.185499] env[63088]: INFO nova.compute.claims [None req-bb8843aa-5c19-4289-a41c-17cbfcb47a00 tempest-ServerActionsTestOtherA-1614359098 tempest-ServerActionsTestOtherA-1614359098-project-member] [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 663.279160] env[63088]: DEBUG nova.network.neutron [None req-442e0a41-8be2-46c5-9922-69b38e5ecc58 tempest-ServerAddressesNegativeTestJSON-62458283 tempest-ServerAddressesNegativeTestJSON-62458283-project-member] [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 663.371175] env[63088]: DEBUG nova.network.neutron [None req-442e0a41-8be2-46c5-9922-69b38e5ecc58 tempest-ServerAddressesNegativeTestJSON-62458283 tempest-ServerAddressesNegativeTestJSON-62458283-project-member] [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 663.689634] env[63088]: DEBUG nova.compute.utils [None req-32cec7a8-18a7-43df-b82f-d466be0dcd4a tempest-AttachInterfacesUnderV243Test-1205880501 tempest-AttachInterfacesUnderV243Test-1205880501-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 663.693662] env[63088]: DEBUG nova.compute.manager [None req-32cec7a8-18a7-43df-b82f-d466be0dcd4a tempest-AttachInterfacesUnderV243Test-1205880501 tempest-AttachInterfacesUnderV243Test-1205880501-project-member] [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 663.693662] env[63088]: DEBUG nova.network.neutron [None req-32cec7a8-18a7-43df-b82f-d466be0dcd4a tempest-AttachInterfacesUnderV243Test-1205880501 tempest-AttachInterfacesUnderV243Test-1205880501-project-member] [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 663.732099] env[63088]: DEBUG nova.policy [None req-32cec7a8-18a7-43df-b82f-d466be0dcd4a tempest-AttachInterfacesUnderV243Test-1205880501 tempest-AttachInterfacesUnderV243Test-1205880501-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f8ebc83a93a048499480acb1dfd91e92', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bce5e6b3cf8d4dcaa5c3440140cb56b3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 663.874560] env[63088]: DEBUG oslo_concurrency.lockutils [None req-442e0a41-8be2-46c5-9922-69b38e5ecc58 tempest-ServerAddressesNegativeTestJSON-62458283 tempest-ServerAddressesNegativeTestJSON-62458283-project-member] Releasing lock "refresh_cache-be5c7b15-c036-4a20-a21b-74bba67c9a2c" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 663.875034] env[63088]: DEBUG nova.compute.manager [None req-442e0a41-8be2-46c5-9922-69b38e5ecc58 tempest-ServerAddressesNegativeTestJSON-62458283 tempest-ServerAddressesNegativeTestJSON-62458283-project-member] [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 663.875865] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-442e0a41-8be2-46c5-9922-69b38e5ecc58 tempest-ServerAddressesNegativeTestJSON-62458283 tempest-ServerAddressesNegativeTestJSON-62458283-project-member] [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 663.875865] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0528912f-e11c-40f0-93a0-5a834c74e3da {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.884883] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aaca2877-a57e-467c-928d-73e41f99cb19 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.907425] env[63088]: WARNING nova.virt.vmwareapi.vmops [None req-442e0a41-8be2-46c5-9922-69b38e5ecc58 tempest-ServerAddressesNegativeTestJSON-62458283 tempest-ServerAddressesNegativeTestJSON-62458283-project-member] [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance be5c7b15-c036-4a20-a21b-74bba67c9a2c could not be found. [ 663.907580] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-442e0a41-8be2-46c5-9922-69b38e5ecc58 tempest-ServerAddressesNegativeTestJSON-62458283 tempest-ServerAddressesNegativeTestJSON-62458283-project-member] [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 663.907765] env[63088]: INFO nova.compute.manager [None req-442e0a41-8be2-46c5-9922-69b38e5ecc58 tempest-ServerAddressesNegativeTestJSON-62458283 tempest-ServerAddressesNegativeTestJSON-62458283-project-member] [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] Took 0.03 seconds to destroy the instance on the hypervisor. [ 663.908036] env[63088]: DEBUG oslo.service.loopingcall [None req-442e0a41-8be2-46c5-9922-69b38e5ecc58 tempest-ServerAddressesNegativeTestJSON-62458283 tempest-ServerAddressesNegativeTestJSON-62458283-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 663.908272] env[63088]: DEBUG nova.compute.manager [-] [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 663.908364] env[63088]: DEBUG nova.network.neutron [-] [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 663.925067] env[63088]: DEBUG nova.network.neutron [-] [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 664.036171] env[63088]: DEBUG nova.compute.manager [req-9cc9464f-7b53-4812-b1f7-a421bb0f3354 req-55a336e8-4ab9-4a03-b8ec-503083121835 service nova] [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] Received event network-changed-be2d3e68-49fa-4eae-8386-e494fa78b11b {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 664.036171] env[63088]: DEBUG nova.compute.manager [req-9cc9464f-7b53-4812-b1f7-a421bb0f3354 req-55a336e8-4ab9-4a03-b8ec-503083121835 service nova] [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] Refreshing instance network info cache due to event network-changed-be2d3e68-49fa-4eae-8386-e494fa78b11b. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 664.036171] env[63088]: DEBUG oslo_concurrency.lockutils [req-9cc9464f-7b53-4812-b1f7-a421bb0f3354 req-55a336e8-4ab9-4a03-b8ec-503083121835 service nova] Acquiring lock "refresh_cache-be5c7b15-c036-4a20-a21b-74bba67c9a2c" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 664.036171] env[63088]: DEBUG oslo_concurrency.lockutils [req-9cc9464f-7b53-4812-b1f7-a421bb0f3354 req-55a336e8-4ab9-4a03-b8ec-503083121835 service nova] Acquired lock "refresh_cache-be5c7b15-c036-4a20-a21b-74bba67c9a2c" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 664.036171] env[63088]: DEBUG nova.network.neutron [req-9cc9464f-7b53-4812-b1f7-a421bb0f3354 req-55a336e8-4ab9-4a03-b8ec-503083121835 service nova] [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] Refreshing network info cache for port be2d3e68-49fa-4eae-8386-e494fa78b11b {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 664.055642] env[63088]: DEBUG nova.network.neutron [None req-32cec7a8-18a7-43df-b82f-d466be0dcd4a tempest-AttachInterfacesUnderV243Test-1205880501 tempest-AttachInterfacesUnderV243Test-1205880501-project-member] [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] Successfully created port: 51e4aed6-7891-4427-90b4-0458a81ef471 {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 664.195778] env[63088]: DEBUG nova.compute.manager [None req-32cec7a8-18a7-43df-b82f-d466be0dcd4a tempest-AttachInterfacesUnderV243Test-1205880501 tempest-AttachInterfacesUnderV243Test-1205880501-project-member] [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 664.427317] env[63088]: DEBUG nova.network.neutron [-] [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 664.564903] env[63088]: DEBUG nova.network.neutron [req-9cc9464f-7b53-4812-b1f7-a421bb0f3354 req-55a336e8-4ab9-4a03-b8ec-503083121835 service nova] [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 664.597023] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb8426c1-6404-45cd-9695-95f4ff5dce21 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.604827] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-943be620-5187-447b-aa12-5f5e90f9a2a6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.640174] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90aa777a-5fb2-4f09-92b1-e20e37dcd2b8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.646540] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c1bfb0a-bc80-4836-ba36-d74617942ac2 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.659639] env[63088]: DEBUG nova.compute.provider_tree [None req-bb8843aa-5c19-4289-a41c-17cbfcb47a00 tempest-ServerActionsTestOtherA-1614359098 tempest-ServerActionsTestOtherA-1614359098-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 664.664019] env[63088]: DEBUG nova.network.neutron [req-9cc9464f-7b53-4812-b1f7-a421bb0f3354 req-55a336e8-4ab9-4a03-b8ec-503083121835 service nova] [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 664.930301] env[63088]: DEBUG nova.compute.manager [req-6027b2e3-0e12-481f-aea9-6c06a6b79fb5 req-c1e68a48-a599-45a3-a167-1a8a9e8f3dd0 service nova] [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] Received event network-changed-51e4aed6-7891-4427-90b4-0458a81ef471 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 664.930301] env[63088]: DEBUG nova.compute.manager [req-6027b2e3-0e12-481f-aea9-6c06a6b79fb5 req-c1e68a48-a599-45a3-a167-1a8a9e8f3dd0 service nova] [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] Refreshing instance network info cache due to event network-changed-51e4aed6-7891-4427-90b4-0458a81ef471. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 664.930301] env[63088]: DEBUG oslo_concurrency.lockutils [req-6027b2e3-0e12-481f-aea9-6c06a6b79fb5 req-c1e68a48-a599-45a3-a167-1a8a9e8f3dd0 service nova] Acquiring lock "refresh_cache-6137d44e-9945-479d-9cf3-4a36bfa0e32c" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 664.930301] env[63088]: DEBUG oslo_concurrency.lockutils [req-6027b2e3-0e12-481f-aea9-6c06a6b79fb5 req-c1e68a48-a599-45a3-a167-1a8a9e8f3dd0 service nova] Acquired lock "refresh_cache-6137d44e-9945-479d-9cf3-4a36bfa0e32c" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 664.930301] env[63088]: DEBUG nova.network.neutron [req-6027b2e3-0e12-481f-aea9-6c06a6b79fb5 req-c1e68a48-a599-45a3-a167-1a8a9e8f3dd0 service nova] [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] Refreshing network info cache for port 51e4aed6-7891-4427-90b4-0458a81ef471 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 664.931160] env[63088]: INFO nova.compute.manager [-] [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] Took 1.02 seconds to deallocate network for instance. [ 664.933296] env[63088]: DEBUG nova.compute.claims [None req-442e0a41-8be2-46c5-9922-69b38e5ecc58 tempest-ServerAddressesNegativeTestJSON-62458283 tempest-ServerAddressesNegativeTestJSON-62458283-project-member] [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] Aborting claim: {{(pid=63088) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 664.933475] env[63088]: DEBUG oslo_concurrency.lockutils [None req-442e0a41-8be2-46c5-9922-69b38e5ecc58 tempest-ServerAddressesNegativeTestJSON-62458283 tempest-ServerAddressesNegativeTestJSON-62458283-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 665.147835] env[63088]: ERROR nova.compute.manager [None req-32cec7a8-18a7-43df-b82f-d466be0dcd4a tempest-AttachInterfacesUnderV243Test-1205880501 tempest-AttachInterfacesUnderV243Test-1205880501-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 51e4aed6-7891-4427-90b4-0458a81ef471, please check neutron logs for more information. [ 665.147835] env[63088]: ERROR nova.compute.manager Traceback (most recent call last): [ 665.147835] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 665.147835] env[63088]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 665.147835] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 665.147835] env[63088]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 665.147835] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 665.147835] env[63088]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 665.147835] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 665.147835] env[63088]: ERROR nova.compute.manager self.force_reraise() [ 665.147835] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 665.147835] env[63088]: ERROR nova.compute.manager raise self.value [ 665.147835] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 665.147835] env[63088]: ERROR nova.compute.manager updated_port = self._update_port( [ 665.147835] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 665.147835] env[63088]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 665.148366] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 665.148366] env[63088]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 665.148366] env[63088]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 51e4aed6-7891-4427-90b4-0458a81ef471, please check neutron logs for more information. [ 665.148366] env[63088]: ERROR nova.compute.manager [ 665.148366] env[63088]: Traceback (most recent call last): [ 665.148366] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 665.148366] env[63088]: listener.cb(fileno) [ 665.148366] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 665.148366] env[63088]: result = function(*args, **kwargs) [ 665.148366] env[63088]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 665.148366] env[63088]: return func(*args, **kwargs) [ 665.148366] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 665.148366] env[63088]: raise e [ 665.148366] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 665.148366] env[63088]: nwinfo = self.network_api.allocate_for_instance( [ 665.148366] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 665.148366] env[63088]: created_port_ids = self._update_ports_for_instance( [ 665.148366] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 665.148366] env[63088]: with excutils.save_and_reraise_exception(): [ 665.148366] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 665.148366] env[63088]: self.force_reraise() [ 665.148366] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 665.148366] env[63088]: raise self.value [ 665.148366] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 665.148366] env[63088]: updated_port = self._update_port( [ 665.148366] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 665.148366] env[63088]: _ensure_no_port_binding_failure(port) [ 665.148366] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 665.148366] env[63088]: raise exception.PortBindingFailed(port_id=port['id']) [ 665.149196] env[63088]: nova.exception.PortBindingFailed: Binding failed for port 51e4aed6-7891-4427-90b4-0458a81ef471, please check neutron logs for more information. [ 665.149196] env[63088]: Removing descriptor: 17 [ 665.164194] env[63088]: DEBUG nova.scheduler.client.report [None req-bb8843aa-5c19-4289-a41c-17cbfcb47a00 tempest-ServerActionsTestOtherA-1614359098 tempest-ServerActionsTestOtherA-1614359098-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 665.167958] env[63088]: DEBUG oslo_concurrency.lockutils [req-9cc9464f-7b53-4812-b1f7-a421bb0f3354 req-55a336e8-4ab9-4a03-b8ec-503083121835 service nova] Releasing lock "refresh_cache-be5c7b15-c036-4a20-a21b-74bba67c9a2c" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 665.167958] env[63088]: DEBUG nova.compute.manager [req-9cc9464f-7b53-4812-b1f7-a421bb0f3354 req-55a336e8-4ab9-4a03-b8ec-503083121835 service nova] [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] Received event network-vif-deleted-be2d3e68-49fa-4eae-8386-e494fa78b11b {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 665.205892] env[63088]: DEBUG nova.compute.manager [None req-32cec7a8-18a7-43df-b82f-d466be0dcd4a tempest-AttachInterfacesUnderV243Test-1205880501 tempest-AttachInterfacesUnderV243Test-1205880501-project-member] [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 665.230973] env[63088]: DEBUG nova.virt.hardware [None req-32cec7a8-18a7-43df-b82f-d466be0dcd4a tempest-AttachInterfacesUnderV243Test-1205880501 tempest-AttachInterfacesUnderV243Test-1205880501-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 665.231291] env[63088]: DEBUG nova.virt.hardware [None req-32cec7a8-18a7-43df-b82f-d466be0dcd4a tempest-AttachInterfacesUnderV243Test-1205880501 tempest-AttachInterfacesUnderV243Test-1205880501-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 665.231491] env[63088]: DEBUG nova.virt.hardware [None req-32cec7a8-18a7-43df-b82f-d466be0dcd4a tempest-AttachInterfacesUnderV243Test-1205880501 tempest-AttachInterfacesUnderV243Test-1205880501-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 665.231720] env[63088]: DEBUG nova.virt.hardware [None req-32cec7a8-18a7-43df-b82f-d466be0dcd4a tempest-AttachInterfacesUnderV243Test-1205880501 tempest-AttachInterfacesUnderV243Test-1205880501-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 665.231939] env[63088]: DEBUG nova.virt.hardware [None req-32cec7a8-18a7-43df-b82f-d466be0dcd4a tempest-AttachInterfacesUnderV243Test-1205880501 tempest-AttachInterfacesUnderV243Test-1205880501-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 665.232139] env[63088]: DEBUG nova.virt.hardware [None req-32cec7a8-18a7-43df-b82f-d466be0dcd4a tempest-AttachInterfacesUnderV243Test-1205880501 tempest-AttachInterfacesUnderV243Test-1205880501-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 665.232384] env[63088]: DEBUG nova.virt.hardware [None req-32cec7a8-18a7-43df-b82f-d466be0dcd4a tempest-AttachInterfacesUnderV243Test-1205880501 tempest-AttachInterfacesUnderV243Test-1205880501-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 665.232579] env[63088]: DEBUG nova.virt.hardware [None req-32cec7a8-18a7-43df-b82f-d466be0dcd4a tempest-AttachInterfacesUnderV243Test-1205880501 tempest-AttachInterfacesUnderV243Test-1205880501-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 665.232797] env[63088]: DEBUG nova.virt.hardware [None req-32cec7a8-18a7-43df-b82f-d466be0dcd4a tempest-AttachInterfacesUnderV243Test-1205880501 tempest-AttachInterfacesUnderV243Test-1205880501-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 665.232969] env[63088]: DEBUG nova.virt.hardware [None req-32cec7a8-18a7-43df-b82f-d466be0dcd4a tempest-AttachInterfacesUnderV243Test-1205880501 tempest-AttachInterfacesUnderV243Test-1205880501-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 665.233276] env[63088]: DEBUG nova.virt.hardware [None req-32cec7a8-18a7-43df-b82f-d466be0dcd4a tempest-AttachInterfacesUnderV243Test-1205880501 tempest-AttachInterfacesUnderV243Test-1205880501-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 665.234167] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-619e427f-77f6-4ca9-aa1c-ec7054ea0a99 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.242844] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-883f251e-bdf6-4d41-8e1a-290de9113b7d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.260669] env[63088]: ERROR nova.compute.manager [None req-32cec7a8-18a7-43df-b82f-d466be0dcd4a tempest-AttachInterfacesUnderV243Test-1205880501 tempest-AttachInterfacesUnderV243Test-1205880501-project-member] [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 51e4aed6-7891-4427-90b4-0458a81ef471, please check neutron logs for more information. [ 665.260669] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] Traceback (most recent call last): [ 665.260669] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 665.260669] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] yield resources [ 665.260669] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 665.260669] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] self.driver.spawn(context, instance, image_meta, [ 665.260669] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 665.260669] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 665.260669] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 665.260669] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] vm_ref = self.build_virtual_machine(instance, [ 665.260669] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 665.261048] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] vif_infos = vmwarevif.get_vif_info(self._session, [ 665.261048] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 665.261048] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] for vif in network_info: [ 665.261048] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 665.261048] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] return self._sync_wrapper(fn, *args, **kwargs) [ 665.261048] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 665.261048] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] self.wait() [ 665.261048] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 665.261048] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] self[:] = self._gt.wait() [ 665.261048] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 665.261048] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] return self._exit_event.wait() [ 665.261048] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 665.261048] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] current.throw(*self._exc) [ 665.261413] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 665.261413] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] result = function(*args, **kwargs) [ 665.261413] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 665.261413] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] return func(*args, **kwargs) [ 665.261413] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 665.261413] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] raise e [ 665.261413] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 665.261413] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] nwinfo = self.network_api.allocate_for_instance( [ 665.261413] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 665.261413] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] created_port_ids = self._update_ports_for_instance( [ 665.261413] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 665.261413] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] with excutils.save_and_reraise_exception(): [ 665.261413] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 665.261790] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] self.force_reraise() [ 665.261790] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 665.261790] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] raise self.value [ 665.261790] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 665.261790] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] updated_port = self._update_port( [ 665.261790] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 665.261790] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] _ensure_no_port_binding_failure(port) [ 665.261790] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 665.261790] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] raise exception.PortBindingFailed(port_id=port['id']) [ 665.261790] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] nova.exception.PortBindingFailed: Binding failed for port 51e4aed6-7891-4427-90b4-0458a81ef471, please check neutron logs for more information. [ 665.261790] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] [ 665.261790] env[63088]: INFO nova.compute.manager [None req-32cec7a8-18a7-43df-b82f-d466be0dcd4a tempest-AttachInterfacesUnderV243Test-1205880501 tempest-AttachInterfacesUnderV243Test-1205880501-project-member] [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] Terminating instance [ 665.263036] env[63088]: DEBUG oslo_concurrency.lockutils [None req-32cec7a8-18a7-43df-b82f-d466be0dcd4a tempest-AttachInterfacesUnderV243Test-1205880501 tempest-AttachInterfacesUnderV243Test-1205880501-project-member] Acquiring lock "refresh_cache-6137d44e-9945-479d-9cf3-4a36bfa0e32c" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 665.445329] env[63088]: DEBUG nova.network.neutron [req-6027b2e3-0e12-481f-aea9-6c06a6b79fb5 req-c1e68a48-a599-45a3-a167-1a8a9e8f3dd0 service nova] [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 665.517978] env[63088]: DEBUG nova.network.neutron [req-6027b2e3-0e12-481f-aea9-6c06a6b79fb5 req-c1e68a48-a599-45a3-a167-1a8a9e8f3dd0 service nova] [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 665.669087] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bb8843aa-5c19-4289-a41c-17cbfcb47a00 tempest-ServerActionsTestOtherA-1614359098 tempest-ServerActionsTestOtherA-1614359098-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.485s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 665.670065] env[63088]: DEBUG nova.compute.manager [None req-bb8843aa-5c19-4289-a41c-17cbfcb47a00 tempest-ServerActionsTestOtherA-1614359098 tempest-ServerActionsTestOtherA-1614359098-project-member] [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 665.672780] env[63088]: DEBUG oslo_concurrency.lockutils [None req-18116a16-f5bf-4db6-aaca-b64d4351d5d4 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.496s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 665.674158] env[63088]: INFO nova.compute.claims [None req-18116a16-f5bf-4db6-aaca-b64d4351d5d4 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 666.020880] env[63088]: DEBUG oslo_concurrency.lockutils [req-6027b2e3-0e12-481f-aea9-6c06a6b79fb5 req-c1e68a48-a599-45a3-a167-1a8a9e8f3dd0 service nova] Releasing lock "refresh_cache-6137d44e-9945-479d-9cf3-4a36bfa0e32c" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 666.021333] env[63088]: DEBUG oslo_concurrency.lockutils [None req-32cec7a8-18a7-43df-b82f-d466be0dcd4a tempest-AttachInterfacesUnderV243Test-1205880501 tempest-AttachInterfacesUnderV243Test-1205880501-project-member] Acquired lock "refresh_cache-6137d44e-9945-479d-9cf3-4a36bfa0e32c" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 666.021522] env[63088]: DEBUG nova.network.neutron [None req-32cec7a8-18a7-43df-b82f-d466be0dcd4a tempest-AttachInterfacesUnderV243Test-1205880501 tempest-AttachInterfacesUnderV243Test-1205880501-project-member] [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 666.180912] env[63088]: DEBUG nova.compute.utils [None req-bb8843aa-5c19-4289-a41c-17cbfcb47a00 tempest-ServerActionsTestOtherA-1614359098 tempest-ServerActionsTestOtherA-1614359098-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 666.183313] env[63088]: DEBUG nova.compute.manager [None req-bb8843aa-5c19-4289-a41c-17cbfcb47a00 tempest-ServerActionsTestOtherA-1614359098 tempest-ServerActionsTestOtherA-1614359098-project-member] [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 666.183490] env[63088]: DEBUG nova.network.neutron [None req-bb8843aa-5c19-4289-a41c-17cbfcb47a00 tempest-ServerActionsTestOtherA-1614359098 tempest-ServerActionsTestOtherA-1614359098-project-member] [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 666.230139] env[63088]: DEBUG nova.policy [None req-bb8843aa-5c19-4289-a41c-17cbfcb47a00 tempest-ServerActionsTestOtherA-1614359098 tempest-ServerActionsTestOtherA-1614359098-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c772b383f54b4e0d81ad170a9ebf456a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7a9adf302f4a4ee48e3e7ddd3709c4ac', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 666.524630] env[63088]: DEBUG nova.network.neutron [None req-bb8843aa-5c19-4289-a41c-17cbfcb47a00 tempest-ServerActionsTestOtherA-1614359098 tempest-ServerActionsTestOtherA-1614359098-project-member] [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] Successfully created port: d9f05fa0-ab0e-479e-ad2f-09a2ece1af2d {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 666.541331] env[63088]: DEBUG nova.network.neutron [None req-32cec7a8-18a7-43df-b82f-d466be0dcd4a tempest-AttachInterfacesUnderV243Test-1205880501 tempest-AttachInterfacesUnderV243Test-1205880501-project-member] [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 666.678447] env[63088]: DEBUG nova.network.neutron [None req-32cec7a8-18a7-43df-b82f-d466be0dcd4a tempest-AttachInterfacesUnderV243Test-1205880501 tempest-AttachInterfacesUnderV243Test-1205880501-project-member] [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 666.687656] env[63088]: DEBUG nova.compute.manager [None req-bb8843aa-5c19-4289-a41c-17cbfcb47a00 tempest-ServerActionsTestOtherA-1614359098 tempest-ServerActionsTestOtherA-1614359098-project-member] [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 666.955903] env[63088]: DEBUG nova.compute.manager [req-ade732e5-712d-47a3-9689-bffec7d1b6d4 req-78e27949-9adb-494e-84ae-ec0859b20fca service nova] [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] Received event network-vif-deleted-51e4aed6-7891-4427-90b4-0458a81ef471 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 667.047201] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e934593-c93e-4795-911d-a3a34f761cd2 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.056226] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7a9d6fa-7a4f-44a2-9952-cbf48dbb83a7 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.089030] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bca8fa7-b832-45ea-9f32-3e2f84225bc7 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.097430] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc360b91-ccc0-4b60-a086-f1590b0fbc99 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.116861] env[63088]: DEBUG nova.compute.provider_tree [None req-18116a16-f5bf-4db6-aaca-b64d4351d5d4 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 667.181414] env[63088]: DEBUG oslo_concurrency.lockutils [None req-32cec7a8-18a7-43df-b82f-d466be0dcd4a tempest-AttachInterfacesUnderV243Test-1205880501 tempest-AttachInterfacesUnderV243Test-1205880501-project-member] Releasing lock "refresh_cache-6137d44e-9945-479d-9cf3-4a36bfa0e32c" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 667.181847] env[63088]: DEBUG nova.compute.manager [None req-32cec7a8-18a7-43df-b82f-d466be0dcd4a tempest-AttachInterfacesUnderV243Test-1205880501 tempest-AttachInterfacesUnderV243Test-1205880501-project-member] [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 667.182187] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-32cec7a8-18a7-43df-b82f-d466be0dcd4a tempest-AttachInterfacesUnderV243Test-1205880501 tempest-AttachInterfacesUnderV243Test-1205880501-project-member] [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 667.182393] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-048021f3-3c20-4d0f-b082-cf5b658b6fd8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.191855] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b4f8a39-2101-49f9-9eb9-5af317798725 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.221025] env[63088]: WARNING nova.virt.vmwareapi.vmops [None req-32cec7a8-18a7-43df-b82f-d466be0dcd4a tempest-AttachInterfacesUnderV243Test-1205880501 tempest-AttachInterfacesUnderV243Test-1205880501-project-member] [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 6137d44e-9945-479d-9cf3-4a36bfa0e32c could not be found. [ 667.221254] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-32cec7a8-18a7-43df-b82f-d466be0dcd4a tempest-AttachInterfacesUnderV243Test-1205880501 tempest-AttachInterfacesUnderV243Test-1205880501-project-member] [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 667.221435] env[63088]: INFO nova.compute.manager [None req-32cec7a8-18a7-43df-b82f-d466be0dcd4a tempest-AttachInterfacesUnderV243Test-1205880501 tempest-AttachInterfacesUnderV243Test-1205880501-project-member] [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] Took 0.04 seconds to destroy the instance on the hypervisor. [ 667.221678] env[63088]: DEBUG oslo.service.loopingcall [None req-32cec7a8-18a7-43df-b82f-d466be0dcd4a tempest-AttachInterfacesUnderV243Test-1205880501 tempest-AttachInterfacesUnderV243Test-1205880501-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 667.221905] env[63088]: DEBUG nova.compute.manager [-] [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 667.221997] env[63088]: DEBUG nova.network.neutron [-] [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 667.238354] env[63088]: DEBUG nova.network.neutron [-] [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 667.499489] env[63088]: ERROR nova.compute.manager [None req-bb8843aa-5c19-4289-a41c-17cbfcb47a00 tempest-ServerActionsTestOtherA-1614359098 tempest-ServerActionsTestOtherA-1614359098-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d9f05fa0-ab0e-479e-ad2f-09a2ece1af2d, please check neutron logs for more information. [ 667.499489] env[63088]: ERROR nova.compute.manager Traceback (most recent call last): [ 667.499489] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 667.499489] env[63088]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 667.499489] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 667.499489] env[63088]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 667.499489] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 667.499489] env[63088]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 667.499489] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 667.499489] env[63088]: ERROR nova.compute.manager self.force_reraise() [ 667.499489] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 667.499489] env[63088]: ERROR nova.compute.manager raise self.value [ 667.499489] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 667.499489] env[63088]: ERROR nova.compute.manager updated_port = self._update_port( [ 667.499489] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 667.499489] env[63088]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 667.500054] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 667.500054] env[63088]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 667.500054] env[63088]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d9f05fa0-ab0e-479e-ad2f-09a2ece1af2d, please check neutron logs for more information. [ 667.500054] env[63088]: ERROR nova.compute.manager [ 667.500054] env[63088]: Traceback (most recent call last): [ 667.500054] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 667.500054] env[63088]: listener.cb(fileno) [ 667.500054] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 667.500054] env[63088]: result = function(*args, **kwargs) [ 667.500054] env[63088]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 667.500054] env[63088]: return func(*args, **kwargs) [ 667.500054] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 667.500054] env[63088]: raise e [ 667.500054] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 667.500054] env[63088]: nwinfo = self.network_api.allocate_for_instance( [ 667.500054] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 667.500054] env[63088]: created_port_ids = self._update_ports_for_instance( [ 667.500054] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 667.500054] env[63088]: with excutils.save_and_reraise_exception(): [ 667.500054] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 667.500054] env[63088]: self.force_reraise() [ 667.500054] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 667.500054] env[63088]: raise self.value [ 667.500054] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 667.500054] env[63088]: updated_port = self._update_port( [ 667.500054] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 667.500054] env[63088]: _ensure_no_port_binding_failure(port) [ 667.500054] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 667.500054] env[63088]: raise exception.PortBindingFailed(port_id=port['id']) [ 667.501192] env[63088]: nova.exception.PortBindingFailed: Binding failed for port d9f05fa0-ab0e-479e-ad2f-09a2ece1af2d, please check neutron logs for more information. [ 667.501192] env[63088]: Removing descriptor: 17 [ 667.621508] env[63088]: DEBUG nova.scheduler.client.report [None req-18116a16-f5bf-4db6-aaca-b64d4351d5d4 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 667.713029] env[63088]: DEBUG nova.compute.manager [None req-bb8843aa-5c19-4289-a41c-17cbfcb47a00 tempest-ServerActionsTestOtherA-1614359098 tempest-ServerActionsTestOtherA-1614359098-project-member] [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 667.743168] env[63088]: DEBUG nova.virt.hardware [None req-bb8843aa-5c19-4289-a41c-17cbfcb47a00 tempest-ServerActionsTestOtherA-1614359098 tempest-ServerActionsTestOtherA-1614359098-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 667.743312] env[63088]: DEBUG nova.virt.hardware [None req-bb8843aa-5c19-4289-a41c-17cbfcb47a00 tempest-ServerActionsTestOtherA-1614359098 tempest-ServerActionsTestOtherA-1614359098-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 667.743419] env[63088]: DEBUG nova.virt.hardware [None req-bb8843aa-5c19-4289-a41c-17cbfcb47a00 tempest-ServerActionsTestOtherA-1614359098 tempest-ServerActionsTestOtherA-1614359098-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 667.743602] env[63088]: DEBUG nova.virt.hardware [None req-bb8843aa-5c19-4289-a41c-17cbfcb47a00 tempest-ServerActionsTestOtherA-1614359098 tempest-ServerActionsTestOtherA-1614359098-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 667.743783] env[63088]: DEBUG nova.virt.hardware [None req-bb8843aa-5c19-4289-a41c-17cbfcb47a00 tempest-ServerActionsTestOtherA-1614359098 tempest-ServerActionsTestOtherA-1614359098-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 667.743923] env[63088]: DEBUG nova.virt.hardware [None req-bb8843aa-5c19-4289-a41c-17cbfcb47a00 tempest-ServerActionsTestOtherA-1614359098 tempest-ServerActionsTestOtherA-1614359098-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 667.744145] env[63088]: DEBUG nova.virt.hardware [None req-bb8843aa-5c19-4289-a41c-17cbfcb47a00 tempest-ServerActionsTestOtherA-1614359098 tempest-ServerActionsTestOtherA-1614359098-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 667.744301] env[63088]: DEBUG nova.virt.hardware [None req-bb8843aa-5c19-4289-a41c-17cbfcb47a00 tempest-ServerActionsTestOtherA-1614359098 tempest-ServerActionsTestOtherA-1614359098-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 667.744512] env[63088]: DEBUG nova.virt.hardware [None req-bb8843aa-5c19-4289-a41c-17cbfcb47a00 tempest-ServerActionsTestOtherA-1614359098 tempest-ServerActionsTestOtherA-1614359098-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 667.744618] env[63088]: DEBUG nova.virt.hardware [None req-bb8843aa-5c19-4289-a41c-17cbfcb47a00 tempest-ServerActionsTestOtherA-1614359098 tempest-ServerActionsTestOtherA-1614359098-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 667.744782] env[63088]: DEBUG nova.virt.hardware [None req-bb8843aa-5c19-4289-a41c-17cbfcb47a00 tempest-ServerActionsTestOtherA-1614359098 tempest-ServerActionsTestOtherA-1614359098-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 667.745114] env[63088]: DEBUG nova.network.neutron [-] [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 667.746763] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04bac725-3236-4b77-a1c5-275537043fb2 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.755683] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b3a1e6d-02d0-448a-8984-128cf9a973b8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.770085] env[63088]: ERROR nova.compute.manager [None req-bb8843aa-5c19-4289-a41c-17cbfcb47a00 tempest-ServerActionsTestOtherA-1614359098 tempest-ServerActionsTestOtherA-1614359098-project-member] [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d9f05fa0-ab0e-479e-ad2f-09a2ece1af2d, please check neutron logs for more information. [ 667.770085] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] Traceback (most recent call last): [ 667.770085] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 667.770085] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] yield resources [ 667.770085] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 667.770085] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] self.driver.spawn(context, instance, image_meta, [ 667.770085] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 667.770085] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 667.770085] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 667.770085] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] vm_ref = self.build_virtual_machine(instance, [ 667.770085] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 667.770513] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] vif_infos = vmwarevif.get_vif_info(self._session, [ 667.770513] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 667.770513] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] for vif in network_info: [ 667.770513] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 667.770513] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] return self._sync_wrapper(fn, *args, **kwargs) [ 667.770513] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 667.770513] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] self.wait() [ 667.770513] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 667.770513] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] self[:] = self._gt.wait() [ 667.770513] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 667.770513] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] return self._exit_event.wait() [ 667.770513] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 667.770513] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] current.throw(*self._exc) [ 667.770816] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 667.770816] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] result = function(*args, **kwargs) [ 667.770816] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 667.770816] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] return func(*args, **kwargs) [ 667.770816] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 667.770816] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] raise e [ 667.770816] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 667.770816] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] nwinfo = self.network_api.allocate_for_instance( [ 667.770816] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 667.770816] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] created_port_ids = self._update_ports_for_instance( [ 667.770816] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 667.770816] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] with excutils.save_and_reraise_exception(): [ 667.770816] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 667.771160] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] self.force_reraise() [ 667.771160] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 667.771160] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] raise self.value [ 667.771160] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 667.771160] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] updated_port = self._update_port( [ 667.771160] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 667.771160] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] _ensure_no_port_binding_failure(port) [ 667.771160] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 667.771160] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] raise exception.PortBindingFailed(port_id=port['id']) [ 667.771160] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] nova.exception.PortBindingFailed: Binding failed for port d9f05fa0-ab0e-479e-ad2f-09a2ece1af2d, please check neutron logs for more information. [ 667.771160] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] [ 667.771160] env[63088]: INFO nova.compute.manager [None req-bb8843aa-5c19-4289-a41c-17cbfcb47a00 tempest-ServerActionsTestOtherA-1614359098 tempest-ServerActionsTestOtherA-1614359098-project-member] [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] Terminating instance [ 667.772246] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bb8843aa-5c19-4289-a41c-17cbfcb47a00 tempest-ServerActionsTestOtherA-1614359098 tempest-ServerActionsTestOtherA-1614359098-project-member] Acquiring lock "refresh_cache-90c1832a-47c5-4f09-9921-21e15cb6fbb6" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 667.772402] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bb8843aa-5c19-4289-a41c-17cbfcb47a00 tempest-ServerActionsTestOtherA-1614359098 tempest-ServerActionsTestOtherA-1614359098-project-member] Acquired lock "refresh_cache-90c1832a-47c5-4f09-9921-21e15cb6fbb6" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 667.772567] env[63088]: DEBUG nova.network.neutron [None req-bb8843aa-5c19-4289-a41c-17cbfcb47a00 tempest-ServerActionsTestOtherA-1614359098 tempest-ServerActionsTestOtherA-1614359098-project-member] [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 668.127087] env[63088]: DEBUG oslo_concurrency.lockutils [None req-18116a16-f5bf-4db6-aaca-b64d4351d5d4 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.454s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 668.127622] env[63088]: DEBUG nova.compute.manager [None req-18116a16-f5bf-4db6-aaca-b64d4351d5d4 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 668.130202] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a5e283fa-bcc1-4186-a4e3-1ae9f8d221af tempest-ServerMetadataTestJSON-859565163 tempest-ServerMetadataTestJSON-859565163-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.754s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 668.251111] env[63088]: INFO nova.compute.manager [-] [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] Took 1.03 seconds to deallocate network for instance. [ 668.253329] env[63088]: DEBUG nova.compute.claims [None req-32cec7a8-18a7-43df-b82f-d466be0dcd4a tempest-AttachInterfacesUnderV243Test-1205880501 tempest-AttachInterfacesUnderV243Test-1205880501-project-member] [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] Aborting claim: {{(pid=63088) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 668.253501] env[63088]: DEBUG oslo_concurrency.lockutils [None req-32cec7a8-18a7-43df-b82f-d466be0dcd4a tempest-AttachInterfacesUnderV243Test-1205880501 tempest-AttachInterfacesUnderV243Test-1205880501-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 668.294487] env[63088]: DEBUG nova.network.neutron [None req-bb8843aa-5c19-4289-a41c-17cbfcb47a00 tempest-ServerActionsTestOtherA-1614359098 tempest-ServerActionsTestOtherA-1614359098-project-member] [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 668.373009] env[63088]: DEBUG nova.network.neutron [None req-bb8843aa-5c19-4289-a41c-17cbfcb47a00 tempest-ServerActionsTestOtherA-1614359098 tempest-ServerActionsTestOtherA-1614359098-project-member] [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 668.636398] env[63088]: DEBUG nova.compute.utils [None req-18116a16-f5bf-4db6-aaca-b64d4351d5d4 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 668.643441] env[63088]: DEBUG nova.compute.manager [None req-18116a16-f5bf-4db6-aaca-b64d4351d5d4 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 668.643441] env[63088]: DEBUG nova.network.neutron [None req-18116a16-f5bf-4db6-aaca-b64d4351d5d4 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 668.680680] env[63088]: DEBUG nova.policy [None req-18116a16-f5bf-4db6-aaca-b64d4351d5d4 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a01468bc41a84e8e8a52f49b6c9a0494', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8e0a144caecd490ebc144409af902624', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 668.877525] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bb8843aa-5c19-4289-a41c-17cbfcb47a00 tempest-ServerActionsTestOtherA-1614359098 tempest-ServerActionsTestOtherA-1614359098-project-member] Releasing lock "refresh_cache-90c1832a-47c5-4f09-9921-21e15cb6fbb6" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 668.877525] env[63088]: DEBUG nova.compute.manager [None req-bb8843aa-5c19-4289-a41c-17cbfcb47a00 tempest-ServerActionsTestOtherA-1614359098 tempest-ServerActionsTestOtherA-1614359098-project-member] [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 668.877525] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-bb8843aa-5c19-4289-a41c-17cbfcb47a00 tempest-ServerActionsTestOtherA-1614359098 tempest-ServerActionsTestOtherA-1614359098-project-member] [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 668.877897] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cdc7169e-6649-4e72-a515-6acc25336ad8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.893598] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64b529f4-9c9c-4aab-a126-4e9a8c6ee90d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.920612] env[63088]: WARNING nova.virt.vmwareapi.vmops [None req-bb8843aa-5c19-4289-a41c-17cbfcb47a00 tempest-ServerActionsTestOtherA-1614359098 tempest-ServerActionsTestOtherA-1614359098-project-member] [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 90c1832a-47c5-4f09-9921-21e15cb6fbb6 could not be found. [ 668.920839] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-bb8843aa-5c19-4289-a41c-17cbfcb47a00 tempest-ServerActionsTestOtherA-1614359098 tempest-ServerActionsTestOtherA-1614359098-project-member] [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 668.921035] env[63088]: INFO nova.compute.manager [None req-bb8843aa-5c19-4289-a41c-17cbfcb47a00 tempest-ServerActionsTestOtherA-1614359098 tempest-ServerActionsTestOtherA-1614359098-project-member] [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] Took 0.04 seconds to destroy the instance on the hypervisor. [ 668.921287] env[63088]: DEBUG oslo.service.loopingcall [None req-bb8843aa-5c19-4289-a41c-17cbfcb47a00 tempest-ServerActionsTestOtherA-1614359098 tempest-ServerActionsTestOtherA-1614359098-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 668.923658] env[63088]: DEBUG nova.compute.manager [-] [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 668.923756] env[63088]: DEBUG nova.network.neutron [-] [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 668.943802] env[63088]: DEBUG nova.network.neutron [-] [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 668.978034] env[63088]: DEBUG nova.network.neutron [None req-18116a16-f5bf-4db6-aaca-b64d4351d5d4 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] Successfully created port: 29c388e4-0159-4c3d-8f7d-9a0cd5ae3d04 {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 668.984942] env[63088]: DEBUG nova.compute.manager [req-bb39df17-a40e-453f-b7c0-45bcd7343e91 req-6e1c4944-bc71-4353-bc2d-f1f9f36d0229 service nova] [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] Received event network-changed-d9f05fa0-ab0e-479e-ad2f-09a2ece1af2d {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 668.985169] env[63088]: DEBUG nova.compute.manager [req-bb39df17-a40e-453f-b7c0-45bcd7343e91 req-6e1c4944-bc71-4353-bc2d-f1f9f36d0229 service nova] [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] Refreshing instance network info cache due to event network-changed-d9f05fa0-ab0e-479e-ad2f-09a2ece1af2d. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 668.985374] env[63088]: DEBUG oslo_concurrency.lockutils [req-bb39df17-a40e-453f-b7c0-45bcd7343e91 req-6e1c4944-bc71-4353-bc2d-f1f9f36d0229 service nova] Acquiring lock "refresh_cache-90c1832a-47c5-4f09-9921-21e15cb6fbb6" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 668.985514] env[63088]: DEBUG oslo_concurrency.lockutils [req-bb39df17-a40e-453f-b7c0-45bcd7343e91 req-6e1c4944-bc71-4353-bc2d-f1f9f36d0229 service nova] Acquired lock "refresh_cache-90c1832a-47c5-4f09-9921-21e15cb6fbb6" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 668.985668] env[63088]: DEBUG nova.network.neutron [req-bb39df17-a40e-453f-b7c0-45bcd7343e91 req-6e1c4944-bc71-4353-bc2d-f1f9f36d0229 service nova] [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] Refreshing network info cache for port d9f05fa0-ab0e-479e-ad2f-09a2ece1af2d {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 669.015259] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3698af6-cf2a-4c55-86d8-92ef6f89be39 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.022908] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c08df7f-6489-4d57-8ea0-06e8eac9e1b2 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.051755] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe1ed026-344e-4216-b47c-c97b001b75ff {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.058941] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce2618ed-a0b5-49b1-bca4-940862f981ff {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.073609] env[63088]: DEBUG nova.compute.provider_tree [None req-a5e283fa-bcc1-4186-a4e3-1ae9f8d221af tempest-ServerMetadataTestJSON-859565163 tempest-ServerMetadataTestJSON-859565163-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 669.141781] env[63088]: DEBUG nova.compute.manager [None req-18116a16-f5bf-4db6-aaca-b64d4351d5d4 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 669.450139] env[63088]: DEBUG nova.network.neutron [-] [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 669.506092] env[63088]: DEBUG nova.network.neutron [req-bb39df17-a40e-453f-b7c0-45bcd7343e91 req-6e1c4944-bc71-4353-bc2d-f1f9f36d0229 service nova] [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 669.576681] env[63088]: DEBUG nova.scheduler.client.report [None req-a5e283fa-bcc1-4186-a4e3-1ae9f8d221af tempest-ServerMetadataTestJSON-859565163 tempest-ServerMetadataTestJSON-859565163-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 669.587642] env[63088]: DEBUG nova.network.neutron [req-bb39df17-a40e-453f-b7c0-45bcd7343e91 req-6e1c4944-bc71-4353-bc2d-f1f9f36d0229 service nova] [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 669.864853] env[63088]: ERROR nova.compute.manager [None req-18116a16-f5bf-4db6-aaca-b64d4351d5d4 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 29c388e4-0159-4c3d-8f7d-9a0cd5ae3d04, please check neutron logs for more information. [ 669.864853] env[63088]: ERROR nova.compute.manager Traceback (most recent call last): [ 669.864853] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 669.864853] env[63088]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 669.864853] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 669.864853] env[63088]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 669.864853] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 669.864853] env[63088]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 669.864853] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 669.864853] env[63088]: ERROR nova.compute.manager self.force_reraise() [ 669.864853] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 669.864853] env[63088]: ERROR nova.compute.manager raise self.value [ 669.864853] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 669.864853] env[63088]: ERROR nova.compute.manager updated_port = self._update_port( [ 669.864853] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 669.864853] env[63088]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 669.865321] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 669.865321] env[63088]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 669.865321] env[63088]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 29c388e4-0159-4c3d-8f7d-9a0cd5ae3d04, please check neutron logs for more information. [ 669.865321] env[63088]: ERROR nova.compute.manager [ 669.865321] env[63088]: Traceback (most recent call last): [ 669.865321] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 669.865321] env[63088]: listener.cb(fileno) [ 669.865321] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 669.865321] env[63088]: result = function(*args, **kwargs) [ 669.865321] env[63088]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 669.865321] env[63088]: return func(*args, **kwargs) [ 669.865321] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 669.865321] env[63088]: raise e [ 669.865321] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 669.865321] env[63088]: nwinfo = self.network_api.allocate_for_instance( [ 669.865321] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 669.865321] env[63088]: created_port_ids = self._update_ports_for_instance( [ 669.865321] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 669.865321] env[63088]: with excutils.save_and_reraise_exception(): [ 669.865321] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 669.865321] env[63088]: self.force_reraise() [ 669.865321] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 669.865321] env[63088]: raise self.value [ 669.865321] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 669.865321] env[63088]: updated_port = self._update_port( [ 669.865321] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 669.865321] env[63088]: _ensure_no_port_binding_failure(port) [ 669.865321] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 669.865321] env[63088]: raise exception.PortBindingFailed(port_id=port['id']) [ 669.865995] env[63088]: nova.exception.PortBindingFailed: Binding failed for port 29c388e4-0159-4c3d-8f7d-9a0cd5ae3d04, please check neutron logs for more information. [ 669.865995] env[63088]: Removing descriptor: 17 [ 669.953040] env[63088]: INFO nova.compute.manager [-] [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] Took 1.03 seconds to deallocate network for instance. [ 669.956774] env[63088]: DEBUG nova.compute.claims [None req-bb8843aa-5c19-4289-a41c-17cbfcb47a00 tempest-ServerActionsTestOtherA-1614359098 tempest-ServerActionsTestOtherA-1614359098-project-member] [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] Aborting claim: {{(pid=63088) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 669.956952] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bb8843aa-5c19-4289-a41c-17cbfcb47a00 tempest-ServerActionsTestOtherA-1614359098 tempest-ServerActionsTestOtherA-1614359098-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 670.086197] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a5e283fa-bcc1-4186-a4e3-1ae9f8d221af tempest-ServerMetadataTestJSON-859565163 tempest-ServerMetadataTestJSON-859565163-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.956s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 670.086922] env[63088]: ERROR nova.compute.manager [None req-a5e283fa-bcc1-4186-a4e3-1ae9f8d221af tempest-ServerMetadataTestJSON-859565163 tempest-ServerMetadataTestJSON-859565163-project-member] [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3aed79cc-3d41-4762-992b-05bb123fc735, please check neutron logs for more information. [ 670.086922] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] Traceback (most recent call last): [ 670.086922] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 670.086922] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] self.driver.spawn(context, instance, image_meta, [ 670.086922] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 670.086922] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 670.086922] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 670.086922] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] vm_ref = self.build_virtual_machine(instance, [ 670.086922] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 670.086922] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] vif_infos = vmwarevif.get_vif_info(self._session, [ 670.086922] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 670.087251] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] for vif in network_info: [ 670.087251] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 670.087251] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] return self._sync_wrapper(fn, *args, **kwargs) [ 670.087251] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 670.087251] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] self.wait() [ 670.087251] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 670.087251] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] self[:] = self._gt.wait() [ 670.087251] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 670.087251] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] return self._exit_event.wait() [ 670.087251] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 670.087251] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] result = hub.switch() [ 670.087251] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 670.087251] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] return self.greenlet.switch() [ 670.087609] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 670.087609] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] result = function(*args, **kwargs) [ 670.087609] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 670.087609] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] return func(*args, **kwargs) [ 670.087609] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 670.087609] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] raise e [ 670.087609] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 670.087609] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] nwinfo = self.network_api.allocate_for_instance( [ 670.087609] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 670.087609] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] created_port_ids = self._update_ports_for_instance( [ 670.087609] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 670.087609] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] with excutils.save_and_reraise_exception(): [ 670.087609] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 670.087986] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] self.force_reraise() [ 670.087986] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 670.087986] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] raise self.value [ 670.087986] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 670.087986] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] updated_port = self._update_port( [ 670.087986] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 670.087986] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] _ensure_no_port_binding_failure(port) [ 670.087986] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 670.087986] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] raise exception.PortBindingFailed(port_id=port['id']) [ 670.087986] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] nova.exception.PortBindingFailed: Binding failed for port 3aed79cc-3d41-4762-992b-05bb123fc735, please check neutron logs for more information. [ 670.087986] env[63088]: ERROR nova.compute.manager [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] [ 670.088304] env[63088]: DEBUG nova.compute.utils [None req-a5e283fa-bcc1-4186-a4e3-1ae9f8d221af tempest-ServerMetadataTestJSON-859565163 tempest-ServerMetadataTestJSON-859565163-project-member] [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] Binding failed for port 3aed79cc-3d41-4762-992b-05bb123fc735, please check neutron logs for more information. {{(pid=63088) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 670.089048] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 17.619s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 670.089223] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 670.089373] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63088) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 670.089650] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f237740d-6eaf-4423-b770-dd9775fe6fd5 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.861s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 670.091085] env[63088]: INFO nova.compute.claims [None req-f237740d-6eaf-4423-b770-dd9775fe6fd5 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 670.093607] env[63088]: DEBUG nova.compute.manager [None req-a5e283fa-bcc1-4186-a4e3-1ae9f8d221af tempest-ServerMetadataTestJSON-859565163 tempest-ServerMetadataTestJSON-859565163-project-member] [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] Build of instance 966fe303-09b9-449f-8bda-4f6cda135f7a was re-scheduled: Binding failed for port 3aed79cc-3d41-4762-992b-05bb123fc735, please check neutron logs for more information. {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 670.094040] env[63088]: DEBUG nova.compute.manager [None req-a5e283fa-bcc1-4186-a4e3-1ae9f8d221af tempest-ServerMetadataTestJSON-859565163 tempest-ServerMetadataTestJSON-859565163-project-member] [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] Unplugging VIFs for instance {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 670.094261] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a5e283fa-bcc1-4186-a4e3-1ae9f8d221af tempest-ServerMetadataTestJSON-859565163 tempest-ServerMetadataTestJSON-859565163-project-member] Acquiring lock "refresh_cache-966fe303-09b9-449f-8bda-4f6cda135f7a" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 670.094404] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a5e283fa-bcc1-4186-a4e3-1ae9f8d221af tempest-ServerMetadataTestJSON-859565163 tempest-ServerMetadataTestJSON-859565163-project-member] Acquired lock "refresh_cache-966fe303-09b9-449f-8bda-4f6cda135f7a" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 670.094560] env[63088]: DEBUG nova.network.neutron [None req-a5e283fa-bcc1-4186-a4e3-1ae9f8d221af tempest-ServerMetadataTestJSON-859565163 tempest-ServerMetadataTestJSON-859565163-project-member] [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 670.095542] env[63088]: DEBUG oslo_concurrency.lockutils [req-bb39df17-a40e-453f-b7c0-45bcd7343e91 req-6e1c4944-bc71-4353-bc2d-f1f9f36d0229 service nova] Releasing lock "refresh_cache-90c1832a-47c5-4f09-9921-21e15cb6fbb6" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 670.095756] env[63088]: DEBUG nova.compute.manager [req-bb39df17-a40e-453f-b7c0-45bcd7343e91 req-6e1c4944-bc71-4353-bc2d-f1f9f36d0229 service nova] [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] Received event network-vif-deleted-d9f05fa0-ab0e-479e-ad2f-09a2ece1af2d {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 670.096628] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6df7cc0-0cb2-4e97-a2d7-b7a8364a6e02 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.105421] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48a05abd-4aff-44fe-b190-451fac7adab2 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.120439] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75e333b9-4f87-48d7-850d-0b63fb5c7912 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.127099] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c2ef1ad-3e01-48ec-a37d-fbf4b49ec35d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.158807] env[63088]: DEBUG nova.compute.manager [None req-18116a16-f5bf-4db6-aaca-b64d4351d5d4 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 670.160951] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181409MB free_disk=140GB free_vcpus=48 pci_devices=None {{(pid=63088) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 670.161834] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 670.186769] env[63088]: DEBUG nova.virt.hardware [None req-18116a16-f5bf-4db6-aaca-b64d4351d5d4 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 670.187514] env[63088]: DEBUG nova.virt.hardware [None req-18116a16-f5bf-4db6-aaca-b64d4351d5d4 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 670.187514] env[63088]: DEBUG nova.virt.hardware [None req-18116a16-f5bf-4db6-aaca-b64d4351d5d4 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 670.187514] env[63088]: DEBUG nova.virt.hardware [None req-18116a16-f5bf-4db6-aaca-b64d4351d5d4 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 670.187651] env[63088]: DEBUG nova.virt.hardware [None req-18116a16-f5bf-4db6-aaca-b64d4351d5d4 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 670.187715] env[63088]: DEBUG nova.virt.hardware [None req-18116a16-f5bf-4db6-aaca-b64d4351d5d4 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 670.187920] env[63088]: DEBUG nova.virt.hardware [None req-18116a16-f5bf-4db6-aaca-b64d4351d5d4 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 670.188102] env[63088]: DEBUG nova.virt.hardware [None req-18116a16-f5bf-4db6-aaca-b64d4351d5d4 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 670.188277] env[63088]: DEBUG nova.virt.hardware [None req-18116a16-f5bf-4db6-aaca-b64d4351d5d4 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 670.188446] env[63088]: DEBUG nova.virt.hardware [None req-18116a16-f5bf-4db6-aaca-b64d4351d5d4 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 670.188619] env[63088]: DEBUG nova.virt.hardware [None req-18116a16-f5bf-4db6-aaca-b64d4351d5d4 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 670.189840] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-589569e4-132d-401b-9cf8-24d645ebf963 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.198684] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94b163cc-7224-47bb-a6f8-098a82615580 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.214151] env[63088]: ERROR nova.compute.manager [None req-18116a16-f5bf-4db6-aaca-b64d4351d5d4 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 29c388e4-0159-4c3d-8f7d-9a0cd5ae3d04, please check neutron logs for more information. [ 670.214151] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] Traceback (most recent call last): [ 670.214151] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 670.214151] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] yield resources [ 670.214151] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 670.214151] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] self.driver.spawn(context, instance, image_meta, [ 670.214151] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 670.214151] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 670.214151] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 670.214151] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] vm_ref = self.build_virtual_machine(instance, [ 670.214151] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 670.214537] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] vif_infos = vmwarevif.get_vif_info(self._session, [ 670.214537] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 670.214537] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] for vif in network_info: [ 670.214537] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 670.214537] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] return self._sync_wrapper(fn, *args, **kwargs) [ 670.214537] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 670.214537] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] self.wait() [ 670.214537] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 670.214537] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] self[:] = self._gt.wait() [ 670.214537] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 670.214537] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] return self._exit_event.wait() [ 670.214537] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 670.214537] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] current.throw(*self._exc) [ 670.215081] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 670.215081] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] result = function(*args, **kwargs) [ 670.215081] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 670.215081] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] return func(*args, **kwargs) [ 670.215081] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 670.215081] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] raise e [ 670.215081] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 670.215081] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] nwinfo = self.network_api.allocate_for_instance( [ 670.215081] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 670.215081] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] created_port_ids = self._update_ports_for_instance( [ 670.215081] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 670.215081] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] with excutils.save_and_reraise_exception(): [ 670.215081] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 670.215475] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] self.force_reraise() [ 670.215475] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 670.215475] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] raise self.value [ 670.215475] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 670.215475] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] updated_port = self._update_port( [ 670.215475] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 670.215475] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] _ensure_no_port_binding_failure(port) [ 670.215475] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 670.215475] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] raise exception.PortBindingFailed(port_id=port['id']) [ 670.215475] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] nova.exception.PortBindingFailed: Binding failed for port 29c388e4-0159-4c3d-8f7d-9a0cd5ae3d04, please check neutron logs for more information. [ 670.215475] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] [ 670.215475] env[63088]: INFO nova.compute.manager [None req-18116a16-f5bf-4db6-aaca-b64d4351d5d4 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] Terminating instance [ 670.216511] env[63088]: DEBUG oslo_concurrency.lockutils [None req-18116a16-f5bf-4db6-aaca-b64d4351d5d4 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Acquiring lock "refresh_cache-0f515940-fa10-4eb8-970d-ead4b996021c" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 670.216684] env[63088]: DEBUG oslo_concurrency.lockutils [None req-18116a16-f5bf-4db6-aaca-b64d4351d5d4 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Acquired lock "refresh_cache-0f515940-fa10-4eb8-970d-ead4b996021c" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 670.216892] env[63088]: DEBUG nova.network.neutron [None req-18116a16-f5bf-4db6-aaca-b64d4351d5d4 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 670.627617] env[63088]: DEBUG nova.network.neutron [None req-a5e283fa-bcc1-4186-a4e3-1ae9f8d221af tempest-ServerMetadataTestJSON-859565163 tempest-ServerMetadataTestJSON-859565163-project-member] [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 670.709341] env[63088]: DEBUG nova.network.neutron [None req-a5e283fa-bcc1-4186-a4e3-1ae9f8d221af tempest-ServerMetadataTestJSON-859565163 tempest-ServerMetadataTestJSON-859565163-project-member] [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 670.733866] env[63088]: DEBUG nova.network.neutron [None req-18116a16-f5bf-4db6-aaca-b64d4351d5d4 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 670.798389] env[63088]: DEBUG nova.network.neutron [None req-18116a16-f5bf-4db6-aaca-b64d4351d5d4 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 671.011587] env[63088]: DEBUG nova.compute.manager [req-d6b08a78-b0f1-4ca7-81b7-13d7ec72b2fd req-70e1c56b-0ce0-4e6d-a142-bf7123bc9a0b service nova] [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] Received event network-changed-29c388e4-0159-4c3d-8f7d-9a0cd5ae3d04 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 671.011587] env[63088]: DEBUG nova.compute.manager [req-d6b08a78-b0f1-4ca7-81b7-13d7ec72b2fd req-70e1c56b-0ce0-4e6d-a142-bf7123bc9a0b service nova] [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] Refreshing instance network info cache due to event network-changed-29c388e4-0159-4c3d-8f7d-9a0cd5ae3d04. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 671.011587] env[63088]: DEBUG oslo_concurrency.lockutils [req-d6b08a78-b0f1-4ca7-81b7-13d7ec72b2fd req-70e1c56b-0ce0-4e6d-a142-bf7123bc9a0b service nova] Acquiring lock "refresh_cache-0f515940-fa10-4eb8-970d-ead4b996021c" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 671.211954] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a5e283fa-bcc1-4186-a4e3-1ae9f8d221af tempest-ServerMetadataTestJSON-859565163 tempest-ServerMetadataTestJSON-859565163-project-member] Releasing lock "refresh_cache-966fe303-09b9-449f-8bda-4f6cda135f7a" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 671.212255] env[63088]: DEBUG nova.compute.manager [None req-a5e283fa-bcc1-4186-a4e3-1ae9f8d221af tempest-ServerMetadataTestJSON-859565163 tempest-ServerMetadataTestJSON-859565163-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 671.212429] env[63088]: DEBUG nova.compute.manager [None req-a5e283fa-bcc1-4186-a4e3-1ae9f8d221af tempest-ServerMetadataTestJSON-859565163 tempest-ServerMetadataTestJSON-859565163-project-member] [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 671.212531] env[63088]: DEBUG nova.network.neutron [None req-a5e283fa-bcc1-4186-a4e3-1ae9f8d221af tempest-ServerMetadataTestJSON-859565163 tempest-ServerMetadataTestJSON-859565163-project-member] [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 671.226679] env[63088]: DEBUG nova.network.neutron [None req-a5e283fa-bcc1-4186-a4e3-1ae9f8d221af tempest-ServerMetadataTestJSON-859565163 tempest-ServerMetadataTestJSON-859565163-project-member] [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 671.301545] env[63088]: DEBUG oslo_concurrency.lockutils [None req-18116a16-f5bf-4db6-aaca-b64d4351d5d4 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Releasing lock "refresh_cache-0f515940-fa10-4eb8-970d-ead4b996021c" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 671.301973] env[63088]: DEBUG nova.compute.manager [None req-18116a16-f5bf-4db6-aaca-b64d4351d5d4 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 671.302182] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-18116a16-f5bf-4db6-aaca-b64d4351d5d4 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 671.302484] env[63088]: DEBUG oslo_concurrency.lockutils [req-d6b08a78-b0f1-4ca7-81b7-13d7ec72b2fd req-70e1c56b-0ce0-4e6d-a142-bf7123bc9a0b service nova] Acquired lock "refresh_cache-0f515940-fa10-4eb8-970d-ead4b996021c" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 671.302659] env[63088]: DEBUG nova.network.neutron [req-d6b08a78-b0f1-4ca7-81b7-13d7ec72b2fd req-70e1c56b-0ce0-4e6d-a142-bf7123bc9a0b service nova] [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] Refreshing network info cache for port 29c388e4-0159-4c3d-8f7d-9a0cd5ae3d04 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 671.303690] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c0a9a811-0ee7-4539-b6d9-d71e9ce21981 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.312392] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b97bc14-56de-4beb-8cc2-bf3c9a5c56b5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.336996] env[63088]: WARNING nova.virt.vmwareapi.vmops [None req-18116a16-f5bf-4db6-aaca-b64d4351d5d4 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0f515940-fa10-4eb8-970d-ead4b996021c could not be found. [ 671.337239] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-18116a16-f5bf-4db6-aaca-b64d4351d5d4 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 671.337438] env[63088]: INFO nova.compute.manager [None req-18116a16-f5bf-4db6-aaca-b64d4351d5d4 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] Took 0.04 seconds to destroy the instance on the hypervisor. [ 671.337685] env[63088]: DEBUG oslo.service.loopingcall [None req-18116a16-f5bf-4db6-aaca-b64d4351d5d4 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 671.337907] env[63088]: DEBUG nova.compute.manager [-] [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 671.338008] env[63088]: DEBUG nova.network.neutron [-] [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 671.353917] env[63088]: DEBUG nova.network.neutron [-] [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 671.512839] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-370a7bd0-22c9-4939-b748-870596858ac1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.521735] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b03efbfc-73f1-4b3f-b465-5e1a1a025b00 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.552086] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74067b09-2a95-4fb7-9f22-49ea34feec21 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.559241] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a35217c7-0a8a-42d0-86f7-8b345619bd2c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.572036] env[63088]: DEBUG nova.compute.provider_tree [None req-f237740d-6eaf-4423-b770-dd9775fe6fd5 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 671.729970] env[63088]: DEBUG nova.network.neutron [None req-a5e283fa-bcc1-4186-a4e3-1ae9f8d221af tempest-ServerMetadataTestJSON-859565163 tempest-ServerMetadataTestJSON-859565163-project-member] [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 671.823185] env[63088]: DEBUG nova.network.neutron [req-d6b08a78-b0f1-4ca7-81b7-13d7ec72b2fd req-70e1c56b-0ce0-4e6d-a142-bf7123bc9a0b service nova] [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 671.856373] env[63088]: DEBUG nova.network.neutron [-] [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 671.896694] env[63088]: DEBUG nova.network.neutron [req-d6b08a78-b0f1-4ca7-81b7-13d7ec72b2fd req-70e1c56b-0ce0-4e6d-a142-bf7123bc9a0b service nova] [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 672.075471] env[63088]: DEBUG nova.scheduler.client.report [None req-f237740d-6eaf-4423-b770-dd9775fe6fd5 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 672.231890] env[63088]: INFO nova.compute.manager [None req-a5e283fa-bcc1-4186-a4e3-1ae9f8d221af tempest-ServerMetadataTestJSON-859565163 tempest-ServerMetadataTestJSON-859565163-project-member] [instance: 966fe303-09b9-449f-8bda-4f6cda135f7a] Took 1.02 seconds to deallocate network for instance. [ 672.358865] env[63088]: INFO nova.compute.manager [-] [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] Took 1.02 seconds to deallocate network for instance. [ 672.362914] env[63088]: DEBUG nova.compute.claims [None req-18116a16-f5bf-4db6-aaca-b64d4351d5d4 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] Aborting claim: {{(pid=63088) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 672.363161] env[63088]: DEBUG oslo_concurrency.lockutils [None req-18116a16-f5bf-4db6-aaca-b64d4351d5d4 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 672.399800] env[63088]: DEBUG oslo_concurrency.lockutils [req-d6b08a78-b0f1-4ca7-81b7-13d7ec72b2fd req-70e1c56b-0ce0-4e6d-a142-bf7123bc9a0b service nova] Releasing lock "refresh_cache-0f515940-fa10-4eb8-970d-ead4b996021c" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 672.400126] env[63088]: DEBUG nova.compute.manager [req-d6b08a78-b0f1-4ca7-81b7-13d7ec72b2fd req-70e1c56b-0ce0-4e6d-a142-bf7123bc9a0b service nova] [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] Received event network-vif-deleted-29c388e4-0159-4c3d-8f7d-9a0cd5ae3d04 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 672.580886] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f237740d-6eaf-4423-b770-dd9775fe6fd5 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.490s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 672.580886] env[63088]: DEBUG nova.compute.manager [None req-f237740d-6eaf-4423-b770-dd9775fe6fd5 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 672.583023] env[63088]: DEBUG oslo_concurrency.lockutils [None req-393edcfb-0595-4190-8ce0-4c14157dbbff tempest-ServerDiagnosticsNegativeTest-1508062492 tempest-ServerDiagnosticsNegativeTest-1508062492-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.289s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 672.584363] env[63088]: INFO nova.compute.claims [None req-393edcfb-0595-4190-8ce0-4c14157dbbff tempest-ServerDiagnosticsNegativeTest-1508062492 tempest-ServerDiagnosticsNegativeTest-1508062492-project-member] [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 673.091501] env[63088]: DEBUG nova.compute.utils [None req-f237740d-6eaf-4423-b770-dd9775fe6fd5 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 673.094584] env[63088]: DEBUG nova.compute.manager [None req-f237740d-6eaf-4423-b770-dd9775fe6fd5 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 673.094745] env[63088]: DEBUG nova.network.neutron [None req-f237740d-6eaf-4423-b770-dd9775fe6fd5 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 673.147301] env[63088]: DEBUG nova.policy [None req-f237740d-6eaf-4423-b770-dd9775fe6fd5 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a01468bc41a84e8e8a52f49b6c9a0494', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8e0a144caecd490ebc144409af902624', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 673.259331] env[63088]: INFO nova.scheduler.client.report [None req-a5e283fa-bcc1-4186-a4e3-1ae9f8d221af tempest-ServerMetadataTestJSON-859565163 tempest-ServerMetadataTestJSON-859565163-project-member] Deleted allocations for instance 966fe303-09b9-449f-8bda-4f6cda135f7a [ 673.413092] env[63088]: DEBUG nova.network.neutron [None req-f237740d-6eaf-4423-b770-dd9775fe6fd5 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] Successfully created port: e454d679-a450-4e9d-81d6-228f3fd67f14 {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 673.600413] env[63088]: DEBUG nova.compute.manager [None req-f237740d-6eaf-4423-b770-dd9775fe6fd5 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 673.771415] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a5e283fa-bcc1-4186-a4e3-1ae9f8d221af tempest-ServerMetadataTestJSON-859565163 tempest-ServerMetadataTestJSON-859565163-project-member] Lock "966fe303-09b9-449f-8bda-4f6cda135f7a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 91.964s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 673.964678] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac31e7b7-a81b-4621-90f8-179959b5ea7b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.972887] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f05e8cb0-6491-4cb0-bf47-e3316f368d71 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.003271] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0660d24-88ee-4684-93dd-9f60b0726732 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.010996] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db43748d-42fc-4850-95ec-29b7b32ea9d7 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.025754] env[63088]: DEBUG nova.compute.provider_tree [None req-393edcfb-0595-4190-8ce0-4c14157dbbff tempest-ServerDiagnosticsNegativeTest-1508062492 tempest-ServerDiagnosticsNegativeTest-1508062492-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 674.226819] env[63088]: DEBUG nova.compute.manager [req-68706950-299c-42b8-aa6f-ac5b3db83e31 req-52ed05b5-97ea-44aa-b0b3-34cce77d44f8 service nova] [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] Received event network-changed-e454d679-a450-4e9d-81d6-228f3fd67f14 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 674.227238] env[63088]: DEBUG nova.compute.manager [req-68706950-299c-42b8-aa6f-ac5b3db83e31 req-52ed05b5-97ea-44aa-b0b3-34cce77d44f8 service nova] [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] Refreshing instance network info cache due to event network-changed-e454d679-a450-4e9d-81d6-228f3fd67f14. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 674.227238] env[63088]: DEBUG oslo_concurrency.lockutils [req-68706950-299c-42b8-aa6f-ac5b3db83e31 req-52ed05b5-97ea-44aa-b0b3-34cce77d44f8 service nova] Acquiring lock "refresh_cache-d9a4d198-4aee-4882-bee8-f6cde14745a2" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 674.227368] env[63088]: DEBUG oslo_concurrency.lockutils [req-68706950-299c-42b8-aa6f-ac5b3db83e31 req-52ed05b5-97ea-44aa-b0b3-34cce77d44f8 service nova] Acquired lock "refresh_cache-d9a4d198-4aee-4882-bee8-f6cde14745a2" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 674.227559] env[63088]: DEBUG nova.network.neutron [req-68706950-299c-42b8-aa6f-ac5b3db83e31 req-52ed05b5-97ea-44aa-b0b3-34cce77d44f8 service nova] [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] Refreshing network info cache for port e454d679-a450-4e9d-81d6-228f3fd67f14 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 674.272703] env[63088]: DEBUG nova.compute.manager [None req-0b29042d-57c3-47a9-a926-f67d685a4740 tempest-TenantUsagesTestJSON-1054626739 tempest-TenantUsagesTestJSON-1054626739-project-member] [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 674.386479] env[63088]: ERROR nova.compute.manager [None req-f237740d-6eaf-4423-b770-dd9775fe6fd5 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e454d679-a450-4e9d-81d6-228f3fd67f14, please check neutron logs for more information. [ 674.386479] env[63088]: ERROR nova.compute.manager Traceback (most recent call last): [ 674.386479] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 674.386479] env[63088]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 674.386479] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 674.386479] env[63088]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 674.386479] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 674.386479] env[63088]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 674.386479] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 674.386479] env[63088]: ERROR nova.compute.manager self.force_reraise() [ 674.386479] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 674.386479] env[63088]: ERROR nova.compute.manager raise self.value [ 674.386479] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 674.386479] env[63088]: ERROR nova.compute.manager updated_port = self._update_port( [ 674.386479] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 674.386479] env[63088]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 674.387058] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 674.387058] env[63088]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 674.387058] env[63088]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e454d679-a450-4e9d-81d6-228f3fd67f14, please check neutron logs for more information. [ 674.387058] env[63088]: ERROR nova.compute.manager [ 674.387058] env[63088]: Traceback (most recent call last): [ 674.387058] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 674.387058] env[63088]: listener.cb(fileno) [ 674.387058] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 674.387058] env[63088]: result = function(*args, **kwargs) [ 674.387058] env[63088]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 674.387058] env[63088]: return func(*args, **kwargs) [ 674.387058] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 674.387058] env[63088]: raise e [ 674.387058] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 674.387058] env[63088]: nwinfo = self.network_api.allocate_for_instance( [ 674.387058] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 674.387058] env[63088]: created_port_ids = self._update_ports_for_instance( [ 674.387058] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 674.387058] env[63088]: with excutils.save_and_reraise_exception(): [ 674.387058] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 674.387058] env[63088]: self.force_reraise() [ 674.387058] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 674.387058] env[63088]: raise self.value [ 674.387058] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 674.387058] env[63088]: updated_port = self._update_port( [ 674.387058] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 674.387058] env[63088]: _ensure_no_port_binding_failure(port) [ 674.387058] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 674.387058] env[63088]: raise exception.PortBindingFailed(port_id=port['id']) [ 674.387841] env[63088]: nova.exception.PortBindingFailed: Binding failed for port e454d679-a450-4e9d-81d6-228f3fd67f14, please check neutron logs for more information. [ 674.387841] env[63088]: Removing descriptor: 17 [ 674.531064] env[63088]: DEBUG nova.scheduler.client.report [None req-393edcfb-0595-4190-8ce0-4c14157dbbff tempest-ServerDiagnosticsNegativeTest-1508062492 tempest-ServerDiagnosticsNegativeTest-1508062492-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 674.613975] env[63088]: DEBUG nova.compute.manager [None req-f237740d-6eaf-4423-b770-dd9775fe6fd5 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 674.643082] env[63088]: DEBUG nova.virt.hardware [None req-f237740d-6eaf-4423-b770-dd9775fe6fd5 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 674.643286] env[63088]: DEBUG nova.virt.hardware [None req-f237740d-6eaf-4423-b770-dd9775fe6fd5 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 674.643415] env[63088]: DEBUG nova.virt.hardware [None req-f237740d-6eaf-4423-b770-dd9775fe6fd5 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 674.643597] env[63088]: DEBUG nova.virt.hardware [None req-f237740d-6eaf-4423-b770-dd9775fe6fd5 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 674.643743] env[63088]: DEBUG nova.virt.hardware [None req-f237740d-6eaf-4423-b770-dd9775fe6fd5 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 674.643889] env[63088]: DEBUG nova.virt.hardware [None req-f237740d-6eaf-4423-b770-dd9775fe6fd5 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 674.644108] env[63088]: DEBUG nova.virt.hardware [None req-f237740d-6eaf-4423-b770-dd9775fe6fd5 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 674.644267] env[63088]: DEBUG nova.virt.hardware [None req-f237740d-6eaf-4423-b770-dd9775fe6fd5 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 674.644432] env[63088]: DEBUG nova.virt.hardware [None req-f237740d-6eaf-4423-b770-dd9775fe6fd5 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 674.644590] env[63088]: DEBUG nova.virt.hardware [None req-f237740d-6eaf-4423-b770-dd9775fe6fd5 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 674.644893] env[63088]: DEBUG nova.virt.hardware [None req-f237740d-6eaf-4423-b770-dd9775fe6fd5 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 674.645637] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d692e104-55b4-4b69-bc87-ce261731cdda {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.654042] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6032e9b-e8ff-43e9-8730-7f945ac9a4ba {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.671015] env[63088]: ERROR nova.compute.manager [None req-f237740d-6eaf-4423-b770-dd9775fe6fd5 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e454d679-a450-4e9d-81d6-228f3fd67f14, please check neutron logs for more information. [ 674.671015] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] Traceback (most recent call last): [ 674.671015] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 674.671015] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] yield resources [ 674.671015] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 674.671015] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] self.driver.spawn(context, instance, image_meta, [ 674.671015] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 674.671015] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 674.671015] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 674.671015] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] vm_ref = self.build_virtual_machine(instance, [ 674.671015] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 674.671432] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] vif_infos = vmwarevif.get_vif_info(self._session, [ 674.671432] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 674.671432] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] for vif in network_info: [ 674.671432] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 674.671432] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] return self._sync_wrapper(fn, *args, **kwargs) [ 674.671432] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 674.671432] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] self.wait() [ 674.671432] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 674.671432] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] self[:] = self._gt.wait() [ 674.671432] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 674.671432] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] return self._exit_event.wait() [ 674.671432] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 674.671432] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] current.throw(*self._exc) [ 674.671799] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 674.671799] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] result = function(*args, **kwargs) [ 674.671799] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 674.671799] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] return func(*args, **kwargs) [ 674.671799] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 674.671799] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] raise e [ 674.671799] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 674.671799] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] nwinfo = self.network_api.allocate_for_instance( [ 674.671799] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 674.671799] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] created_port_ids = self._update_ports_for_instance( [ 674.671799] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 674.671799] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] with excutils.save_and_reraise_exception(): [ 674.671799] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 674.672184] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] self.force_reraise() [ 674.672184] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 674.672184] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] raise self.value [ 674.672184] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 674.672184] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] updated_port = self._update_port( [ 674.672184] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 674.672184] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] _ensure_no_port_binding_failure(port) [ 674.672184] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 674.672184] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] raise exception.PortBindingFailed(port_id=port['id']) [ 674.672184] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] nova.exception.PortBindingFailed: Binding failed for port e454d679-a450-4e9d-81d6-228f3fd67f14, please check neutron logs for more information. [ 674.672184] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] [ 674.672184] env[63088]: INFO nova.compute.manager [None req-f237740d-6eaf-4423-b770-dd9775fe6fd5 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] Terminating instance [ 674.672624] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f237740d-6eaf-4423-b770-dd9775fe6fd5 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Acquiring lock "refresh_cache-d9a4d198-4aee-4882-bee8-f6cde14745a2" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 674.795043] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0b29042d-57c3-47a9-a926-f67d685a4740 tempest-TenantUsagesTestJSON-1054626739 tempest-TenantUsagesTestJSON-1054626739-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 674.883250] env[63088]: DEBUG nova.network.neutron [req-68706950-299c-42b8-aa6f-ac5b3db83e31 req-52ed05b5-97ea-44aa-b0b3-34cce77d44f8 service nova] [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 675.001224] env[63088]: DEBUG nova.network.neutron [req-68706950-299c-42b8-aa6f-ac5b3db83e31 req-52ed05b5-97ea-44aa-b0b3-34cce77d44f8 service nova] [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 675.037137] env[63088]: DEBUG oslo_concurrency.lockutils [None req-393edcfb-0595-4190-8ce0-4c14157dbbff tempest-ServerDiagnosticsNegativeTest-1508062492 tempest-ServerDiagnosticsNegativeTest-1508062492-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.454s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 675.037644] env[63088]: DEBUG nova.compute.manager [None req-393edcfb-0595-4190-8ce0-4c14157dbbff tempest-ServerDiagnosticsNegativeTest-1508062492 tempest-ServerDiagnosticsNegativeTest-1508062492-project-member] [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 675.040146] env[63088]: DEBUG oslo_concurrency.lockutils [None req-224c474a-3c21-4a3f-9526-5bb4e3ba7d78 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.648s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 675.503376] env[63088]: DEBUG oslo_concurrency.lockutils [req-68706950-299c-42b8-aa6f-ac5b3db83e31 req-52ed05b5-97ea-44aa-b0b3-34cce77d44f8 service nova] Releasing lock "refresh_cache-d9a4d198-4aee-4882-bee8-f6cde14745a2" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 675.503871] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f237740d-6eaf-4423-b770-dd9775fe6fd5 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Acquired lock "refresh_cache-d9a4d198-4aee-4882-bee8-f6cde14745a2" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 675.503987] env[63088]: DEBUG nova.network.neutron [None req-f237740d-6eaf-4423-b770-dd9775fe6fd5 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 675.548893] env[63088]: DEBUG nova.compute.utils [None req-393edcfb-0595-4190-8ce0-4c14157dbbff tempest-ServerDiagnosticsNegativeTest-1508062492 tempest-ServerDiagnosticsNegativeTest-1508062492-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 675.551061] env[63088]: DEBUG nova.compute.manager [None req-393edcfb-0595-4190-8ce0-4c14157dbbff tempest-ServerDiagnosticsNegativeTest-1508062492 tempest-ServerDiagnosticsNegativeTest-1508062492-project-member] [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 675.551061] env[63088]: DEBUG nova.network.neutron [None req-393edcfb-0595-4190-8ce0-4c14157dbbff tempest-ServerDiagnosticsNegativeTest-1508062492 tempest-ServerDiagnosticsNegativeTest-1508062492-project-member] [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 675.638402] env[63088]: DEBUG nova.policy [None req-393edcfb-0595-4190-8ce0-4c14157dbbff tempest-ServerDiagnosticsNegativeTest-1508062492 tempest-ServerDiagnosticsNegativeTest-1508062492-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b656a8e1085c45ae938aab414b1006c9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3a3b9086d06f45e2a73885efb4dd20b4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 675.924306] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b379e73-f2bb-4807-8c5d-7ab584d3a8f9 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.932125] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18555d91-db4a-4856-b47c-78f7510ab85f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.963237] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7aa8107-0c92-4e10-be53-2ec09eaf10b4 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.970601] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fdacc26-acc2-4d01-8895-85ea8ce2d655 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.983731] env[63088]: DEBUG nova.compute.provider_tree [None req-224c474a-3c21-4a3f-9526-5bb4e3ba7d78 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 676.023262] env[63088]: DEBUG nova.network.neutron [None req-f237740d-6eaf-4423-b770-dd9775fe6fd5 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 676.032580] env[63088]: DEBUG nova.network.neutron [None req-393edcfb-0595-4190-8ce0-4c14157dbbff tempest-ServerDiagnosticsNegativeTest-1508062492 tempest-ServerDiagnosticsNegativeTest-1508062492-project-member] [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] Successfully created port: b4cabb9e-94ed-49f7-ab53-3e736fbbc48a {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 676.054324] env[63088]: DEBUG nova.compute.manager [None req-393edcfb-0595-4190-8ce0-4c14157dbbff tempest-ServerDiagnosticsNegativeTest-1508062492 tempest-ServerDiagnosticsNegativeTest-1508062492-project-member] [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 676.106168] env[63088]: DEBUG nova.network.neutron [None req-f237740d-6eaf-4423-b770-dd9775fe6fd5 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 676.286736] env[63088]: DEBUG nova.compute.manager [req-dbb1e9cd-b0d1-40b8-be2f-3e53579623e5 req-435d2de6-8c11-4045-9e97-b90d3b059700 service nova] [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] Received event network-vif-deleted-e454d679-a450-4e9d-81d6-228f3fd67f14 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 676.488356] env[63088]: DEBUG nova.scheduler.client.report [None req-224c474a-3c21-4a3f-9526-5bb4e3ba7d78 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 676.608347] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f237740d-6eaf-4423-b770-dd9775fe6fd5 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Releasing lock "refresh_cache-d9a4d198-4aee-4882-bee8-f6cde14745a2" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 676.609250] env[63088]: DEBUG nova.compute.manager [None req-f237740d-6eaf-4423-b770-dd9775fe6fd5 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 676.609250] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-f237740d-6eaf-4423-b770-dd9775fe6fd5 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 676.609912] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f2452979-307b-46a8-939c-412e1c4f129b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.619059] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bf44cac-859e-4a04-a2ca-696f7ac0522d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.642877] env[63088]: WARNING nova.virt.vmwareapi.vmops [None req-f237740d-6eaf-4423-b770-dd9775fe6fd5 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d9a4d198-4aee-4882-bee8-f6cde14745a2 could not be found. [ 676.644054] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-f237740d-6eaf-4423-b770-dd9775fe6fd5 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 676.644054] env[63088]: INFO nova.compute.manager [None req-f237740d-6eaf-4423-b770-dd9775fe6fd5 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] Took 0.03 seconds to destroy the instance on the hypervisor. [ 676.644054] env[63088]: DEBUG oslo.service.loopingcall [None req-f237740d-6eaf-4423-b770-dd9775fe6fd5 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 676.644054] env[63088]: DEBUG nova.compute.manager [-] [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 676.644054] env[63088]: DEBUG nova.network.neutron [-] [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 676.664472] env[63088]: DEBUG nova.network.neutron [-] [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 676.960811] env[63088]: ERROR nova.compute.manager [None req-393edcfb-0595-4190-8ce0-4c14157dbbff tempest-ServerDiagnosticsNegativeTest-1508062492 tempest-ServerDiagnosticsNegativeTest-1508062492-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b4cabb9e-94ed-49f7-ab53-3e736fbbc48a, please check neutron logs for more information. [ 676.960811] env[63088]: ERROR nova.compute.manager Traceback (most recent call last): [ 676.960811] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 676.960811] env[63088]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 676.960811] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 676.960811] env[63088]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 676.960811] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 676.960811] env[63088]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 676.960811] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 676.960811] env[63088]: ERROR nova.compute.manager self.force_reraise() [ 676.960811] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 676.960811] env[63088]: ERROR nova.compute.manager raise self.value [ 676.960811] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 676.960811] env[63088]: ERROR nova.compute.manager updated_port = self._update_port( [ 676.960811] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 676.960811] env[63088]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 676.961256] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 676.961256] env[63088]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 676.961256] env[63088]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b4cabb9e-94ed-49f7-ab53-3e736fbbc48a, please check neutron logs for more information. [ 676.961256] env[63088]: ERROR nova.compute.manager [ 676.961451] env[63088]: Traceback (most recent call last): [ 676.961451] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 676.961451] env[63088]: listener.cb(fileno) [ 676.961451] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 676.961451] env[63088]: result = function(*args, **kwargs) [ 676.961451] env[63088]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 676.961451] env[63088]: return func(*args, **kwargs) [ 676.961451] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 676.961451] env[63088]: raise e [ 676.961451] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 676.961451] env[63088]: nwinfo = self.network_api.allocate_for_instance( [ 676.961451] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 676.961451] env[63088]: created_port_ids = self._update_ports_for_instance( [ 676.961451] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 676.961451] env[63088]: with excutils.save_and_reraise_exception(): [ 676.961451] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 676.961451] env[63088]: self.force_reraise() [ 676.961451] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 676.961451] env[63088]: raise self.value [ 676.961451] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 676.961451] env[63088]: updated_port = self._update_port( [ 676.961451] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 676.961451] env[63088]: _ensure_no_port_binding_failure(port) [ 676.961451] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 676.961451] env[63088]: raise exception.PortBindingFailed(port_id=port['id']) [ 676.961451] env[63088]: nova.exception.PortBindingFailed: Binding failed for port b4cabb9e-94ed-49f7-ab53-3e736fbbc48a, please check neutron logs for more information. [ 676.961451] env[63088]: Removing descriptor: 17 [ 676.990825] env[63088]: DEBUG oslo_concurrency.lockutils [None req-224c474a-3c21-4a3f-9526-5bb4e3ba7d78 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.951s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 676.991558] env[63088]: ERROR nova.compute.manager [None req-224c474a-3c21-4a3f-9526-5bb4e3ba7d78 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port bfee4990-260b-40da-bbcf-869c85e6b40f, please check neutron logs for more information. [ 676.991558] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] Traceback (most recent call last): [ 676.991558] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 676.991558] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] self.driver.spawn(context, instance, image_meta, [ 676.991558] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 676.991558] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] self._vmops.spawn(context, instance, image_meta, injected_files, [ 676.991558] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 676.991558] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] vm_ref = self.build_virtual_machine(instance, [ 676.991558] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 676.991558] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] vif_infos = vmwarevif.get_vif_info(self._session, [ 676.991558] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 676.991865] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] for vif in network_info: [ 676.991865] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 676.991865] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] return self._sync_wrapper(fn, *args, **kwargs) [ 676.991865] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 676.991865] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] self.wait() [ 676.991865] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 676.991865] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] self[:] = self._gt.wait() [ 676.991865] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 676.991865] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] return self._exit_event.wait() [ 676.991865] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 676.991865] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] result = hub.switch() [ 676.991865] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 676.991865] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] return self.greenlet.switch() [ 676.992281] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 676.992281] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] result = function(*args, **kwargs) [ 676.992281] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 676.992281] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] return func(*args, **kwargs) [ 676.992281] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 676.992281] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] raise e [ 676.992281] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 676.992281] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] nwinfo = self.network_api.allocate_for_instance( [ 676.992281] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 676.992281] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] created_port_ids = self._update_ports_for_instance( [ 676.992281] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 676.992281] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] with excutils.save_and_reraise_exception(): [ 676.992281] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 676.992587] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] self.force_reraise() [ 676.992587] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 676.992587] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] raise self.value [ 676.992587] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 676.992587] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] updated_port = self._update_port( [ 676.992587] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 676.992587] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] _ensure_no_port_binding_failure(port) [ 676.992587] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 676.992587] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] raise exception.PortBindingFailed(port_id=port['id']) [ 676.992587] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] nova.exception.PortBindingFailed: Binding failed for port bfee4990-260b-40da-bbcf-869c85e6b40f, please check neutron logs for more information. [ 676.992587] env[63088]: ERROR nova.compute.manager [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] [ 676.992845] env[63088]: DEBUG nova.compute.utils [None req-224c474a-3c21-4a3f-9526-5bb4e3ba7d78 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] Binding failed for port bfee4990-260b-40da-bbcf-869c85e6b40f, please check neutron logs for more information. {{(pid=63088) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 676.993592] env[63088]: DEBUG oslo_concurrency.lockutils [None req-cab9e1ae-5fd3-4816-8684-8d11c94f36d2 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.305s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 676.997102] env[63088]: DEBUG nova.compute.manager [None req-224c474a-3c21-4a3f-9526-5bb4e3ba7d78 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] Build of instance 62bd4013-f007-4f69-b037-14bedf1e63cf was re-scheduled: Binding failed for port bfee4990-260b-40da-bbcf-869c85e6b40f, please check neutron logs for more information. {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 676.997547] env[63088]: DEBUG nova.compute.manager [None req-224c474a-3c21-4a3f-9526-5bb4e3ba7d78 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] Unplugging VIFs for instance {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 676.997768] env[63088]: DEBUG oslo_concurrency.lockutils [None req-224c474a-3c21-4a3f-9526-5bb4e3ba7d78 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] Acquiring lock "refresh_cache-62bd4013-f007-4f69-b037-14bedf1e63cf" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 676.997929] env[63088]: DEBUG oslo_concurrency.lockutils [None req-224c474a-3c21-4a3f-9526-5bb4e3ba7d78 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] Acquired lock "refresh_cache-62bd4013-f007-4f69-b037-14bedf1e63cf" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 676.998120] env[63088]: DEBUG nova.network.neutron [None req-224c474a-3c21-4a3f-9526-5bb4e3ba7d78 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 677.065586] env[63088]: DEBUG nova.compute.manager [None req-393edcfb-0595-4190-8ce0-4c14157dbbff tempest-ServerDiagnosticsNegativeTest-1508062492 tempest-ServerDiagnosticsNegativeTest-1508062492-project-member] [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 677.093118] env[63088]: DEBUG nova.virt.hardware [None req-393edcfb-0595-4190-8ce0-4c14157dbbff tempest-ServerDiagnosticsNegativeTest-1508062492 tempest-ServerDiagnosticsNegativeTest-1508062492-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 677.093373] env[63088]: DEBUG nova.virt.hardware [None req-393edcfb-0595-4190-8ce0-4c14157dbbff tempest-ServerDiagnosticsNegativeTest-1508062492 tempest-ServerDiagnosticsNegativeTest-1508062492-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 677.093532] env[63088]: DEBUG nova.virt.hardware [None req-393edcfb-0595-4190-8ce0-4c14157dbbff tempest-ServerDiagnosticsNegativeTest-1508062492 tempest-ServerDiagnosticsNegativeTest-1508062492-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 677.093712] env[63088]: DEBUG nova.virt.hardware [None req-393edcfb-0595-4190-8ce0-4c14157dbbff tempest-ServerDiagnosticsNegativeTest-1508062492 tempest-ServerDiagnosticsNegativeTest-1508062492-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 677.093861] env[63088]: DEBUG nova.virt.hardware [None req-393edcfb-0595-4190-8ce0-4c14157dbbff tempest-ServerDiagnosticsNegativeTest-1508062492 tempest-ServerDiagnosticsNegativeTest-1508062492-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 677.094015] env[63088]: DEBUG nova.virt.hardware [None req-393edcfb-0595-4190-8ce0-4c14157dbbff tempest-ServerDiagnosticsNegativeTest-1508062492 tempest-ServerDiagnosticsNegativeTest-1508062492-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 677.094279] env[63088]: DEBUG nova.virt.hardware [None req-393edcfb-0595-4190-8ce0-4c14157dbbff tempest-ServerDiagnosticsNegativeTest-1508062492 tempest-ServerDiagnosticsNegativeTest-1508062492-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 677.094367] env[63088]: DEBUG nova.virt.hardware [None req-393edcfb-0595-4190-8ce0-4c14157dbbff tempest-ServerDiagnosticsNegativeTest-1508062492 tempest-ServerDiagnosticsNegativeTest-1508062492-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 677.094531] env[63088]: DEBUG nova.virt.hardware [None req-393edcfb-0595-4190-8ce0-4c14157dbbff tempest-ServerDiagnosticsNegativeTest-1508062492 tempest-ServerDiagnosticsNegativeTest-1508062492-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 677.094691] env[63088]: DEBUG nova.virt.hardware [None req-393edcfb-0595-4190-8ce0-4c14157dbbff tempest-ServerDiagnosticsNegativeTest-1508062492 tempest-ServerDiagnosticsNegativeTest-1508062492-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 677.094860] env[63088]: DEBUG nova.virt.hardware [None req-393edcfb-0595-4190-8ce0-4c14157dbbff tempest-ServerDiagnosticsNegativeTest-1508062492 tempest-ServerDiagnosticsNegativeTest-1508062492-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 677.095799] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54729809-30dc-4c9a-9360-c1b0adc76bdf {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.104269] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36297021-5a32-499b-ba09-1aa9582eee4a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.117828] env[63088]: ERROR nova.compute.manager [None req-393edcfb-0595-4190-8ce0-4c14157dbbff tempest-ServerDiagnosticsNegativeTest-1508062492 tempest-ServerDiagnosticsNegativeTest-1508062492-project-member] [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b4cabb9e-94ed-49f7-ab53-3e736fbbc48a, please check neutron logs for more information. [ 677.117828] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] Traceback (most recent call last): [ 677.117828] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 677.117828] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] yield resources [ 677.117828] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 677.117828] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] self.driver.spawn(context, instance, image_meta, [ 677.117828] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 677.117828] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 677.117828] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 677.117828] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] vm_ref = self.build_virtual_machine(instance, [ 677.117828] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 677.118217] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] vif_infos = vmwarevif.get_vif_info(self._session, [ 677.118217] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 677.118217] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] for vif in network_info: [ 677.118217] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 677.118217] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] return self._sync_wrapper(fn, *args, **kwargs) [ 677.118217] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 677.118217] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] self.wait() [ 677.118217] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 677.118217] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] self[:] = self._gt.wait() [ 677.118217] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 677.118217] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] return self._exit_event.wait() [ 677.118217] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 677.118217] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] current.throw(*self._exc) [ 677.118583] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 677.118583] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] result = function(*args, **kwargs) [ 677.118583] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 677.118583] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] return func(*args, **kwargs) [ 677.118583] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 677.118583] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] raise e [ 677.118583] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 677.118583] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] nwinfo = self.network_api.allocate_for_instance( [ 677.118583] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 677.118583] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] created_port_ids = self._update_ports_for_instance( [ 677.118583] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 677.118583] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] with excutils.save_and_reraise_exception(): [ 677.118583] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 677.119012] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] self.force_reraise() [ 677.119012] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 677.119012] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] raise self.value [ 677.119012] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 677.119012] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] updated_port = self._update_port( [ 677.119012] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 677.119012] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] _ensure_no_port_binding_failure(port) [ 677.119012] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 677.119012] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] raise exception.PortBindingFailed(port_id=port['id']) [ 677.119012] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] nova.exception.PortBindingFailed: Binding failed for port b4cabb9e-94ed-49f7-ab53-3e736fbbc48a, please check neutron logs for more information. [ 677.119012] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] [ 677.119012] env[63088]: INFO nova.compute.manager [None req-393edcfb-0595-4190-8ce0-4c14157dbbff tempest-ServerDiagnosticsNegativeTest-1508062492 tempest-ServerDiagnosticsNegativeTest-1508062492-project-member] [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] Terminating instance [ 677.120195] env[63088]: DEBUG oslo_concurrency.lockutils [None req-393edcfb-0595-4190-8ce0-4c14157dbbff tempest-ServerDiagnosticsNegativeTest-1508062492 tempest-ServerDiagnosticsNegativeTest-1508062492-project-member] Acquiring lock "refresh_cache-6eb3e767-ff7d-4760-a036-a55647afac0f" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 677.120351] env[63088]: DEBUG oslo_concurrency.lockutils [None req-393edcfb-0595-4190-8ce0-4c14157dbbff tempest-ServerDiagnosticsNegativeTest-1508062492 tempest-ServerDiagnosticsNegativeTest-1508062492-project-member] Acquired lock "refresh_cache-6eb3e767-ff7d-4760-a036-a55647afac0f" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 677.120518] env[63088]: DEBUG nova.network.neutron [None req-393edcfb-0595-4190-8ce0-4c14157dbbff tempest-ServerDiagnosticsNegativeTest-1508062492 tempest-ServerDiagnosticsNegativeTest-1508062492-project-member] [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 677.170162] env[63088]: DEBUG nova.network.neutron [-] [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 677.518222] env[63088]: DEBUG nova.network.neutron [None req-224c474a-3c21-4a3f-9526-5bb4e3ba7d78 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 677.566935] env[63088]: DEBUG nova.network.neutron [None req-224c474a-3c21-4a3f-9526-5bb4e3ba7d78 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 677.637225] env[63088]: DEBUG nova.network.neutron [None req-393edcfb-0595-4190-8ce0-4c14157dbbff tempest-ServerDiagnosticsNegativeTest-1508062492 tempest-ServerDiagnosticsNegativeTest-1508062492-project-member] [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 677.673263] env[63088]: INFO nova.compute.manager [-] [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] Took 1.03 seconds to deallocate network for instance. [ 677.676239] env[63088]: DEBUG nova.compute.claims [None req-f237740d-6eaf-4423-b770-dd9775fe6fd5 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] Aborting claim: {{(pid=63088) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 677.676430] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f237740d-6eaf-4423-b770-dd9775fe6fd5 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 677.680164] env[63088]: DEBUG nova.network.neutron [None req-393edcfb-0595-4190-8ce0-4c14157dbbff tempest-ServerDiagnosticsNegativeTest-1508062492 tempest-ServerDiagnosticsNegativeTest-1508062492-project-member] [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 677.806672] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87e6757f-5b02-415c-90fd-5f7d3bc00dcb {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.815802] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b90ab5f-8f29-402c-9d34-6972f42d1093 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.847058] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8fba509-8127-4b52-8dcb-457f77412729 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.854479] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ddab478-af7d-475b-a7d6-1280b0759986 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.867362] env[63088]: DEBUG nova.compute.provider_tree [None req-cab9e1ae-5fd3-4816-8684-8d11c94f36d2 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 678.069672] env[63088]: DEBUG oslo_concurrency.lockutils [None req-224c474a-3c21-4a3f-9526-5bb4e3ba7d78 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] Releasing lock "refresh_cache-62bd4013-f007-4f69-b037-14bedf1e63cf" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 678.070054] env[63088]: DEBUG nova.compute.manager [None req-224c474a-3c21-4a3f-9526-5bb4e3ba7d78 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 678.070236] env[63088]: DEBUG nova.compute.manager [None req-224c474a-3c21-4a3f-9526-5bb4e3ba7d78 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 678.070403] env[63088]: DEBUG nova.network.neutron [None req-224c474a-3c21-4a3f-9526-5bb4e3ba7d78 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 678.092309] env[63088]: DEBUG nova.network.neutron [None req-224c474a-3c21-4a3f-9526-5bb4e3ba7d78 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 678.188060] env[63088]: DEBUG oslo_concurrency.lockutils [None req-393edcfb-0595-4190-8ce0-4c14157dbbff tempest-ServerDiagnosticsNegativeTest-1508062492 tempest-ServerDiagnosticsNegativeTest-1508062492-project-member] Releasing lock "refresh_cache-6eb3e767-ff7d-4760-a036-a55647afac0f" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 678.188060] env[63088]: DEBUG nova.compute.manager [None req-393edcfb-0595-4190-8ce0-4c14157dbbff tempest-ServerDiagnosticsNegativeTest-1508062492 tempest-ServerDiagnosticsNegativeTest-1508062492-project-member] [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 678.188060] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-393edcfb-0595-4190-8ce0-4c14157dbbff tempest-ServerDiagnosticsNegativeTest-1508062492 tempest-ServerDiagnosticsNegativeTest-1508062492-project-member] [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 678.188299] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9ae2604a-6781-4d96-bdb7-93f50b46802d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.197009] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efc53f3f-21de-4b86-a0d4-040ef0ff8a3b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.218425] env[63088]: WARNING nova.virt.vmwareapi.vmops [None req-393edcfb-0595-4190-8ce0-4c14157dbbff tempest-ServerDiagnosticsNegativeTest-1508062492 tempest-ServerDiagnosticsNegativeTest-1508062492-project-member] [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 6eb3e767-ff7d-4760-a036-a55647afac0f could not be found. [ 678.218622] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-393edcfb-0595-4190-8ce0-4c14157dbbff tempest-ServerDiagnosticsNegativeTest-1508062492 tempest-ServerDiagnosticsNegativeTest-1508062492-project-member] [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 678.218823] env[63088]: INFO nova.compute.manager [None req-393edcfb-0595-4190-8ce0-4c14157dbbff tempest-ServerDiagnosticsNegativeTest-1508062492 tempest-ServerDiagnosticsNegativeTest-1508062492-project-member] [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] Took 0.03 seconds to destroy the instance on the hypervisor. [ 678.219078] env[63088]: DEBUG oslo.service.loopingcall [None req-393edcfb-0595-4190-8ce0-4c14157dbbff tempest-ServerDiagnosticsNegativeTest-1508062492 tempest-ServerDiagnosticsNegativeTest-1508062492-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 678.219276] env[63088]: DEBUG nova.compute.manager [-] [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 678.219360] env[63088]: DEBUG nova.network.neutron [-] [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 678.234590] env[63088]: DEBUG nova.network.neutron [-] [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 678.337554] env[63088]: DEBUG nova.compute.manager [req-f24c0cac-96b3-427c-a147-a2ddd09dd9c4 req-8892b9be-04df-4d7b-8d0b-3dfd49f35b74 service nova] [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] Received event network-changed-b4cabb9e-94ed-49f7-ab53-3e736fbbc48a {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 678.337686] env[63088]: DEBUG nova.compute.manager [req-f24c0cac-96b3-427c-a147-a2ddd09dd9c4 req-8892b9be-04df-4d7b-8d0b-3dfd49f35b74 service nova] [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] Refreshing instance network info cache due to event network-changed-b4cabb9e-94ed-49f7-ab53-3e736fbbc48a. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 678.337901] env[63088]: DEBUG oslo_concurrency.lockutils [req-f24c0cac-96b3-427c-a147-a2ddd09dd9c4 req-8892b9be-04df-4d7b-8d0b-3dfd49f35b74 service nova] Acquiring lock "refresh_cache-6eb3e767-ff7d-4760-a036-a55647afac0f" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 678.338062] env[63088]: DEBUG oslo_concurrency.lockutils [req-f24c0cac-96b3-427c-a147-a2ddd09dd9c4 req-8892b9be-04df-4d7b-8d0b-3dfd49f35b74 service nova] Acquired lock "refresh_cache-6eb3e767-ff7d-4760-a036-a55647afac0f" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 678.338225] env[63088]: DEBUG nova.network.neutron [req-f24c0cac-96b3-427c-a147-a2ddd09dd9c4 req-8892b9be-04df-4d7b-8d0b-3dfd49f35b74 service nova] [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] Refreshing network info cache for port b4cabb9e-94ed-49f7-ab53-3e736fbbc48a {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 678.370721] env[63088]: DEBUG nova.scheduler.client.report [None req-cab9e1ae-5fd3-4816-8684-8d11c94f36d2 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 678.594755] env[63088]: DEBUG nova.network.neutron [None req-224c474a-3c21-4a3f-9526-5bb4e3ba7d78 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 678.738047] env[63088]: DEBUG nova.network.neutron [-] [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 678.859929] env[63088]: DEBUG nova.network.neutron [req-f24c0cac-96b3-427c-a147-a2ddd09dd9c4 req-8892b9be-04df-4d7b-8d0b-3dfd49f35b74 service nova] [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 678.876564] env[63088]: DEBUG oslo_concurrency.lockutils [None req-cab9e1ae-5fd3-4816-8684-8d11c94f36d2 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.883s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 678.877242] env[63088]: ERROR nova.compute.manager [None req-cab9e1ae-5fd3-4816-8684-8d11c94f36d2 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port eb9c1776-bb3b-4b18-9c46-5dc6e7f78df5, please check neutron logs for more information. [ 678.877242] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] Traceback (most recent call last): [ 678.877242] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 678.877242] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] self.driver.spawn(context, instance, image_meta, [ 678.877242] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 678.877242] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] self._vmops.spawn(context, instance, image_meta, injected_files, [ 678.877242] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 678.877242] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] vm_ref = self.build_virtual_machine(instance, [ 678.877242] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 678.877242] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] vif_infos = vmwarevif.get_vif_info(self._session, [ 678.877242] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 678.877575] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] for vif in network_info: [ 678.877575] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 678.877575] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] return self._sync_wrapper(fn, *args, **kwargs) [ 678.877575] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 678.877575] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] self.wait() [ 678.877575] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 678.877575] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] self[:] = self._gt.wait() [ 678.877575] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 678.877575] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] return self._exit_event.wait() [ 678.877575] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 678.877575] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] result = hub.switch() [ 678.877575] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 678.877575] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] return self.greenlet.switch() [ 678.877936] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 678.877936] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] result = function(*args, **kwargs) [ 678.877936] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 678.877936] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] return func(*args, **kwargs) [ 678.877936] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 678.877936] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] raise e [ 678.877936] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 678.877936] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] nwinfo = self.network_api.allocate_for_instance( [ 678.877936] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 678.877936] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] created_port_ids = self._update_ports_for_instance( [ 678.877936] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 678.877936] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] with excutils.save_and_reraise_exception(): [ 678.877936] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 678.878303] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] self.force_reraise() [ 678.878303] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 678.878303] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] raise self.value [ 678.878303] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 678.878303] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] updated_port = self._update_port( [ 678.878303] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 678.878303] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] _ensure_no_port_binding_failure(port) [ 678.878303] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 678.878303] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] raise exception.PortBindingFailed(port_id=port['id']) [ 678.878303] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] nova.exception.PortBindingFailed: Binding failed for port eb9c1776-bb3b-4b18-9c46-5dc6e7f78df5, please check neutron logs for more information. [ 678.878303] env[63088]: ERROR nova.compute.manager [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] [ 678.878608] env[63088]: DEBUG nova.compute.utils [None req-cab9e1ae-5fd3-4816-8684-8d11c94f36d2 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] Binding failed for port eb9c1776-bb3b-4b18-9c46-5dc6e7f78df5, please check neutron logs for more information. {{(pid=63088) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 678.881047] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b7ea81d3-fe92-4160-8c6b-de9afb157d35 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.481s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 678.883921] env[63088]: DEBUG nova.compute.manager [None req-cab9e1ae-5fd3-4816-8684-8d11c94f36d2 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] Build of instance a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af was re-scheduled: Binding failed for port eb9c1776-bb3b-4b18-9c46-5dc6e7f78df5, please check neutron logs for more information. {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 678.884422] env[63088]: DEBUG nova.compute.manager [None req-cab9e1ae-5fd3-4816-8684-8d11c94f36d2 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] Unplugging VIFs for instance {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 678.884662] env[63088]: DEBUG oslo_concurrency.lockutils [None req-cab9e1ae-5fd3-4816-8684-8d11c94f36d2 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] Acquiring lock "refresh_cache-a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 678.884832] env[63088]: DEBUG oslo_concurrency.lockutils [None req-cab9e1ae-5fd3-4816-8684-8d11c94f36d2 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] Acquired lock "refresh_cache-a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 678.885022] env[63088]: DEBUG nova.network.neutron [None req-cab9e1ae-5fd3-4816-8684-8d11c94f36d2 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 678.912973] env[63088]: DEBUG nova.network.neutron [req-f24c0cac-96b3-427c-a147-a2ddd09dd9c4 req-8892b9be-04df-4d7b-8d0b-3dfd49f35b74 service nova] [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 679.096947] env[63088]: INFO nova.compute.manager [None req-224c474a-3c21-4a3f-9526-5bb4e3ba7d78 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] [instance: 62bd4013-f007-4f69-b037-14bedf1e63cf] Took 1.03 seconds to deallocate network for instance. [ 679.241175] env[63088]: INFO nova.compute.manager [-] [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] Took 1.02 seconds to deallocate network for instance. [ 679.243629] env[63088]: DEBUG nova.compute.claims [None req-393edcfb-0595-4190-8ce0-4c14157dbbff tempest-ServerDiagnosticsNegativeTest-1508062492 tempest-ServerDiagnosticsNegativeTest-1508062492-project-member] [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] Aborting claim: {{(pid=63088) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 679.243865] env[63088]: DEBUG oslo_concurrency.lockutils [None req-393edcfb-0595-4190-8ce0-4c14157dbbff tempest-ServerDiagnosticsNegativeTest-1508062492 tempest-ServerDiagnosticsNegativeTest-1508062492-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 679.405040] env[63088]: DEBUG nova.network.neutron [None req-cab9e1ae-5fd3-4816-8684-8d11c94f36d2 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 679.415804] env[63088]: DEBUG oslo_concurrency.lockutils [req-f24c0cac-96b3-427c-a147-a2ddd09dd9c4 req-8892b9be-04df-4d7b-8d0b-3dfd49f35b74 service nova] Releasing lock "refresh_cache-6eb3e767-ff7d-4760-a036-a55647afac0f" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 679.415898] env[63088]: DEBUG nova.compute.manager [req-f24c0cac-96b3-427c-a147-a2ddd09dd9c4 req-8892b9be-04df-4d7b-8d0b-3dfd49f35b74 service nova] [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] Received event network-vif-deleted-b4cabb9e-94ed-49f7-ab53-3e736fbbc48a {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 679.460035] env[63088]: DEBUG nova.network.neutron [None req-cab9e1ae-5fd3-4816-8684-8d11c94f36d2 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 679.678788] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a13f64d-5c41-43b6-9c10-1f1f6c956167 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.686576] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61f2d30f-0716-4695-85a6-edd1a216d106 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.715952] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79410f4d-8c25-4443-a6a0-c621fc707559 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.723215] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25954ae3-f3c0-4746-927f-8d5e7ba3f00d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.736075] env[63088]: DEBUG nova.compute.provider_tree [None req-b7ea81d3-fe92-4160-8c6b-de9afb157d35 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 679.964129] env[63088]: DEBUG oslo_concurrency.lockutils [None req-cab9e1ae-5fd3-4816-8684-8d11c94f36d2 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] Releasing lock "refresh_cache-a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 679.964409] env[63088]: DEBUG nova.compute.manager [None req-cab9e1ae-5fd3-4816-8684-8d11c94f36d2 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 679.964590] env[63088]: DEBUG nova.compute.manager [None req-cab9e1ae-5fd3-4816-8684-8d11c94f36d2 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 679.964804] env[63088]: DEBUG nova.network.neutron [None req-cab9e1ae-5fd3-4816-8684-8d11c94f36d2 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 679.979270] env[63088]: DEBUG nova.network.neutron [None req-cab9e1ae-5fd3-4816-8684-8d11c94f36d2 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 680.129493] env[63088]: INFO nova.scheduler.client.report [None req-224c474a-3c21-4a3f-9526-5bb4e3ba7d78 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] Deleted allocations for instance 62bd4013-f007-4f69-b037-14bedf1e63cf [ 680.239502] env[63088]: DEBUG nova.scheduler.client.report [None req-b7ea81d3-fe92-4160-8c6b-de9afb157d35 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 680.482241] env[63088]: DEBUG nova.network.neutron [None req-cab9e1ae-5fd3-4816-8684-8d11c94f36d2 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 680.643925] env[63088]: DEBUG oslo_concurrency.lockutils [None req-224c474a-3c21-4a3f-9526-5bb4e3ba7d78 tempest-DeleteServersAdminTestJSON-652739439 tempest-DeleteServersAdminTestJSON-652739439-project-member] Lock "62bd4013-f007-4f69-b037-14bedf1e63cf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 97.959s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 680.745132] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b7ea81d3-fe92-4160-8c6b-de9afb157d35 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.864s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 680.745963] env[63088]: ERROR nova.compute.manager [None req-b7ea81d3-fe92-4160-8c6b-de9afb157d35 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 7c5a6f5a-771f-481d-a6f2-f56e1d85dfc9, please check neutron logs for more information. [ 680.745963] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] Traceback (most recent call last): [ 680.745963] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 680.745963] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] self.driver.spawn(context, instance, image_meta, [ 680.745963] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 680.745963] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] self._vmops.spawn(context, instance, image_meta, injected_files, [ 680.745963] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 680.745963] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] vm_ref = self.build_virtual_machine(instance, [ 680.745963] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 680.745963] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] vif_infos = vmwarevif.get_vif_info(self._session, [ 680.745963] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 680.746547] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] for vif in network_info: [ 680.746547] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 680.746547] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] return self._sync_wrapper(fn, *args, **kwargs) [ 680.746547] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 680.746547] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] self.wait() [ 680.746547] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 680.746547] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] self[:] = self._gt.wait() [ 680.746547] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 680.746547] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] return self._exit_event.wait() [ 680.746547] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 680.746547] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] result = hub.switch() [ 680.746547] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 680.746547] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] return self.greenlet.switch() [ 680.747109] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 680.747109] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] result = function(*args, **kwargs) [ 680.747109] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 680.747109] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] return func(*args, **kwargs) [ 680.747109] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 680.747109] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] raise e [ 680.747109] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 680.747109] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] nwinfo = self.network_api.allocate_for_instance( [ 680.747109] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 680.747109] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] created_port_ids = self._update_ports_for_instance( [ 680.747109] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 680.747109] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] with excutils.save_and_reraise_exception(): [ 680.747109] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 680.747681] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] self.force_reraise() [ 680.747681] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 680.747681] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] raise self.value [ 680.747681] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 680.747681] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] updated_port = self._update_port( [ 680.747681] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 680.747681] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] _ensure_no_port_binding_failure(port) [ 680.747681] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 680.747681] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] raise exception.PortBindingFailed(port_id=port['id']) [ 680.747681] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] nova.exception.PortBindingFailed: Binding failed for port 7c5a6f5a-771f-481d-a6f2-f56e1d85dfc9, please check neutron logs for more information. [ 680.747681] env[63088]: ERROR nova.compute.manager [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] [ 680.748186] env[63088]: DEBUG nova.compute.utils [None req-b7ea81d3-fe92-4160-8c6b-de9afb157d35 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] Binding failed for port 7c5a6f5a-771f-481d-a6f2-f56e1d85dfc9, please check neutron logs for more information. {{(pid=63088) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 680.748640] env[63088]: DEBUG oslo_concurrency.lockutils [None req-442e0a41-8be2-46c5-9922-69b38e5ecc58 tempest-ServerAddressesNegativeTestJSON-62458283 tempest-ServerAddressesNegativeTestJSON-62458283-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.815s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 680.751690] env[63088]: DEBUG nova.compute.manager [None req-b7ea81d3-fe92-4160-8c6b-de9afb157d35 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] Build of instance 945b5a77-96b9-4565-94a8-6f1a89dc3bbf was re-scheduled: Binding failed for port 7c5a6f5a-771f-481d-a6f2-f56e1d85dfc9, please check neutron logs for more information. {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 680.752221] env[63088]: DEBUG nova.compute.manager [None req-b7ea81d3-fe92-4160-8c6b-de9afb157d35 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] Unplugging VIFs for instance {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 680.752499] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b7ea81d3-fe92-4160-8c6b-de9afb157d35 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] Acquiring lock "refresh_cache-945b5a77-96b9-4565-94a8-6f1a89dc3bbf" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 680.752702] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b7ea81d3-fe92-4160-8c6b-de9afb157d35 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] Acquired lock "refresh_cache-945b5a77-96b9-4565-94a8-6f1a89dc3bbf" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 680.752929] env[63088]: DEBUG nova.network.neutron [None req-b7ea81d3-fe92-4160-8c6b-de9afb157d35 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 680.984624] env[63088]: INFO nova.compute.manager [None req-cab9e1ae-5fd3-4816-8684-8d11c94f36d2 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] [instance: a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af] Took 1.02 seconds to deallocate network for instance. [ 681.147261] env[63088]: DEBUG nova.compute.manager [None req-fcd5880e-5b8a-42fa-b638-13cd9836ac96 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 681.276809] env[63088]: DEBUG nova.network.neutron [None req-b7ea81d3-fe92-4160-8c6b-de9afb157d35 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 681.351296] env[63088]: DEBUG nova.network.neutron [None req-b7ea81d3-fe92-4160-8c6b-de9afb157d35 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 681.642742] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3acc097b-5d5f-4149-ba52-6bf5e9aa7ceb {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.650989] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-daa33617-b09f-486a-b293-9acd0e0e21f5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.685992] env[63088]: DEBUG oslo_concurrency.lockutils [None req-fcd5880e-5b8a-42fa-b638-13cd9836ac96 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 681.686872] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a01da57-fb27-4906-8a8a-2c256458c612 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.696202] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8e0d0d1-dcbe-4f85-9337-2d05a90c31f8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.710129] env[63088]: DEBUG nova.compute.provider_tree [None req-442e0a41-8be2-46c5-9922-69b38e5ecc58 tempest-ServerAddressesNegativeTestJSON-62458283 tempest-ServerAddressesNegativeTestJSON-62458283-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 681.854765] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b7ea81d3-fe92-4160-8c6b-de9afb157d35 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] Releasing lock "refresh_cache-945b5a77-96b9-4565-94a8-6f1a89dc3bbf" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 681.855075] env[63088]: DEBUG nova.compute.manager [None req-b7ea81d3-fe92-4160-8c6b-de9afb157d35 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 681.855314] env[63088]: DEBUG nova.compute.manager [None req-b7ea81d3-fe92-4160-8c6b-de9afb157d35 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 681.855519] env[63088]: DEBUG nova.network.neutron [None req-b7ea81d3-fe92-4160-8c6b-de9afb157d35 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 681.874115] env[63088]: DEBUG nova.network.neutron [None req-b7ea81d3-fe92-4160-8c6b-de9afb157d35 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 682.021173] env[63088]: INFO nova.scheduler.client.report [None req-cab9e1ae-5fd3-4816-8684-8d11c94f36d2 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] Deleted allocations for instance a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af [ 682.213165] env[63088]: DEBUG nova.scheduler.client.report [None req-442e0a41-8be2-46c5-9922-69b38e5ecc58 tempest-ServerAddressesNegativeTestJSON-62458283 tempest-ServerAddressesNegativeTestJSON-62458283-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 682.376752] env[63088]: DEBUG nova.network.neutron [None req-b7ea81d3-fe92-4160-8c6b-de9afb157d35 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 682.532098] env[63088]: DEBUG oslo_concurrency.lockutils [None req-cab9e1ae-5fd3-4816-8684-8d11c94f36d2 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] Lock "a1e51c6f-fc0c-4ce0-a695-d6ef2f36d8af" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 99.577s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 682.719944] env[63088]: DEBUG oslo_concurrency.lockutils [None req-442e0a41-8be2-46c5-9922-69b38e5ecc58 tempest-ServerAddressesNegativeTestJSON-62458283 tempest-ServerAddressesNegativeTestJSON-62458283-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.970s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 682.719944] env[63088]: ERROR nova.compute.manager [None req-442e0a41-8be2-46c5-9922-69b38e5ecc58 tempest-ServerAddressesNegativeTestJSON-62458283 tempest-ServerAddressesNegativeTestJSON-62458283-project-member] [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port be2d3e68-49fa-4eae-8386-e494fa78b11b, please check neutron logs for more information. [ 682.719944] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] Traceback (most recent call last): [ 682.719944] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 682.719944] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] self.driver.spawn(context, instance, image_meta, [ 682.719944] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 682.719944] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 682.719944] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 682.719944] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] vm_ref = self.build_virtual_machine(instance, [ 682.720387] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 682.720387] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] vif_infos = vmwarevif.get_vif_info(self._session, [ 682.720387] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 682.720387] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] for vif in network_info: [ 682.720387] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 682.720387] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] return self._sync_wrapper(fn, *args, **kwargs) [ 682.720387] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 682.720387] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] self.wait() [ 682.720387] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 682.720387] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] self[:] = self._gt.wait() [ 682.720387] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 682.720387] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] return self._exit_event.wait() [ 682.720387] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 682.720845] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] current.throw(*self._exc) [ 682.720845] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 682.720845] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] result = function(*args, **kwargs) [ 682.720845] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 682.720845] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] return func(*args, **kwargs) [ 682.720845] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 682.720845] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] raise e [ 682.720845] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 682.720845] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] nwinfo = self.network_api.allocate_for_instance( [ 682.720845] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 682.720845] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] created_port_ids = self._update_ports_for_instance( [ 682.720845] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 682.720845] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] with excutils.save_and_reraise_exception(): [ 682.721206] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 682.721206] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] self.force_reraise() [ 682.721206] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 682.721206] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] raise self.value [ 682.721206] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 682.721206] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] updated_port = self._update_port( [ 682.721206] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 682.721206] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] _ensure_no_port_binding_failure(port) [ 682.721206] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 682.721206] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] raise exception.PortBindingFailed(port_id=port['id']) [ 682.721206] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] nova.exception.PortBindingFailed: Binding failed for port be2d3e68-49fa-4eae-8386-e494fa78b11b, please check neutron logs for more information. [ 682.721206] env[63088]: ERROR nova.compute.manager [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] [ 682.721977] env[63088]: DEBUG nova.compute.utils [None req-442e0a41-8be2-46c5-9922-69b38e5ecc58 tempest-ServerAddressesNegativeTestJSON-62458283 tempest-ServerAddressesNegativeTestJSON-62458283-project-member] [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] Binding failed for port be2d3e68-49fa-4eae-8386-e494fa78b11b, please check neutron logs for more information. {{(pid=63088) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 682.721977] env[63088]: DEBUG oslo_concurrency.lockutils [None req-32cec7a8-18a7-43df-b82f-d466be0dcd4a tempest-AttachInterfacesUnderV243Test-1205880501 tempest-AttachInterfacesUnderV243Test-1205880501-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.468s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 682.725247] env[63088]: DEBUG nova.compute.manager [None req-442e0a41-8be2-46c5-9922-69b38e5ecc58 tempest-ServerAddressesNegativeTestJSON-62458283 tempest-ServerAddressesNegativeTestJSON-62458283-project-member] [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] Build of instance be5c7b15-c036-4a20-a21b-74bba67c9a2c was re-scheduled: Binding failed for port be2d3e68-49fa-4eae-8386-e494fa78b11b, please check neutron logs for more information. {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 682.725684] env[63088]: DEBUG nova.compute.manager [None req-442e0a41-8be2-46c5-9922-69b38e5ecc58 tempest-ServerAddressesNegativeTestJSON-62458283 tempest-ServerAddressesNegativeTestJSON-62458283-project-member] [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] Unplugging VIFs for instance {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 682.725928] env[63088]: DEBUG oslo_concurrency.lockutils [None req-442e0a41-8be2-46c5-9922-69b38e5ecc58 tempest-ServerAddressesNegativeTestJSON-62458283 tempest-ServerAddressesNegativeTestJSON-62458283-project-member] Acquiring lock "refresh_cache-be5c7b15-c036-4a20-a21b-74bba67c9a2c" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 682.726108] env[63088]: DEBUG oslo_concurrency.lockutils [None req-442e0a41-8be2-46c5-9922-69b38e5ecc58 tempest-ServerAddressesNegativeTestJSON-62458283 tempest-ServerAddressesNegativeTestJSON-62458283-project-member] Acquired lock "refresh_cache-be5c7b15-c036-4a20-a21b-74bba67c9a2c" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 682.726277] env[63088]: DEBUG nova.network.neutron [None req-442e0a41-8be2-46c5-9922-69b38e5ecc58 tempest-ServerAddressesNegativeTestJSON-62458283 tempest-ServerAddressesNegativeTestJSON-62458283-project-member] [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 682.879252] env[63088]: INFO nova.compute.manager [None req-b7ea81d3-fe92-4160-8c6b-de9afb157d35 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] [instance: 945b5a77-96b9-4565-94a8-6f1a89dc3bbf] Took 1.02 seconds to deallocate network for instance. [ 683.034986] env[63088]: DEBUG nova.compute.manager [None req-6c63332a-ec25-45cc-9207-42e0be237d32 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 683.246760] env[63088]: DEBUG nova.network.neutron [None req-442e0a41-8be2-46c5-9922-69b38e5ecc58 tempest-ServerAddressesNegativeTestJSON-62458283 tempest-ServerAddressesNegativeTestJSON-62458283-project-member] [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 683.348139] env[63088]: DEBUG nova.network.neutron [None req-442e0a41-8be2-46c5-9922-69b38e5ecc58 tempest-ServerAddressesNegativeTestJSON-62458283 tempest-ServerAddressesNegativeTestJSON-62458283-project-member] [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 683.376465] env[63088]: DEBUG oslo_concurrency.lockutils [None req-02aa52fc-420f-46cd-9829-a68800c429e2 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Acquiring lock "e6a72b08-38b6-44ae-8911-6b39d50ee1c3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 683.377784] env[63088]: DEBUG oslo_concurrency.lockutils [None req-02aa52fc-420f-46cd-9829-a68800c429e2 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Lock "e6a72b08-38b6-44ae-8911-6b39d50ee1c3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 683.554254] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6c63332a-ec25-45cc-9207-42e0be237d32 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 683.573889] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c804039-6d90-4d2d-a130-04a19397fbe2 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.580792] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac65cf14-3684-46ff-ba02-c10242d2020c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.609442] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9741dbbf-ecf2-4035-93ec-e06f75891a8d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.616657] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfed2f75-c01a-4636-9034-a5483be2ea8b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.629087] env[63088]: DEBUG nova.compute.provider_tree [None req-32cec7a8-18a7-43df-b82f-d466be0dcd4a tempest-AttachInterfacesUnderV243Test-1205880501 tempest-AttachInterfacesUnderV243Test-1205880501-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 683.851017] env[63088]: DEBUG oslo_concurrency.lockutils [None req-442e0a41-8be2-46c5-9922-69b38e5ecc58 tempest-ServerAddressesNegativeTestJSON-62458283 tempest-ServerAddressesNegativeTestJSON-62458283-project-member] Releasing lock "refresh_cache-be5c7b15-c036-4a20-a21b-74bba67c9a2c" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 683.851286] env[63088]: DEBUG nova.compute.manager [None req-442e0a41-8be2-46c5-9922-69b38e5ecc58 tempest-ServerAddressesNegativeTestJSON-62458283 tempest-ServerAddressesNegativeTestJSON-62458283-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 683.851471] env[63088]: DEBUG nova.compute.manager [None req-442e0a41-8be2-46c5-9922-69b38e5ecc58 tempest-ServerAddressesNegativeTestJSON-62458283 tempest-ServerAddressesNegativeTestJSON-62458283-project-member] [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 683.851671] env[63088]: DEBUG nova.network.neutron [None req-442e0a41-8be2-46c5-9922-69b38e5ecc58 tempest-ServerAddressesNegativeTestJSON-62458283 tempest-ServerAddressesNegativeTestJSON-62458283-project-member] [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 683.866589] env[63088]: DEBUG nova.network.neutron [None req-442e0a41-8be2-46c5-9922-69b38e5ecc58 tempest-ServerAddressesNegativeTestJSON-62458283 tempest-ServerAddressesNegativeTestJSON-62458283-project-member] [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 683.923767] env[63088]: INFO nova.scheduler.client.report [None req-b7ea81d3-fe92-4160-8c6b-de9afb157d35 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] Deleted allocations for instance 945b5a77-96b9-4565-94a8-6f1a89dc3bbf [ 684.132443] env[63088]: DEBUG nova.scheduler.client.report [None req-32cec7a8-18a7-43df-b82f-d466be0dcd4a tempest-AttachInterfacesUnderV243Test-1205880501 tempest-AttachInterfacesUnderV243Test-1205880501-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 684.369251] env[63088]: DEBUG nova.network.neutron [None req-442e0a41-8be2-46c5-9922-69b38e5ecc58 tempest-ServerAddressesNegativeTestJSON-62458283 tempest-ServerAddressesNegativeTestJSON-62458283-project-member] [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 684.431979] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b7ea81d3-fe92-4160-8c6b-de9afb157d35 tempest-ListImageFiltersTestJSON-255983591 tempest-ListImageFiltersTestJSON-255983591-project-member] Lock "945b5a77-96b9-4565-94a8-6f1a89dc3bbf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 100.097s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 684.640511] env[63088]: DEBUG oslo_concurrency.lockutils [None req-32cec7a8-18a7-43df-b82f-d466be0dcd4a tempest-AttachInterfacesUnderV243Test-1205880501 tempest-AttachInterfacesUnderV243Test-1205880501-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.919s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 684.641174] env[63088]: ERROR nova.compute.manager [None req-32cec7a8-18a7-43df-b82f-d466be0dcd4a tempest-AttachInterfacesUnderV243Test-1205880501 tempest-AttachInterfacesUnderV243Test-1205880501-project-member] [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 51e4aed6-7891-4427-90b4-0458a81ef471, please check neutron logs for more information. [ 684.641174] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] Traceback (most recent call last): [ 684.641174] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 684.641174] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] self.driver.spawn(context, instance, image_meta, [ 684.641174] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 684.641174] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 684.641174] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 684.641174] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] vm_ref = self.build_virtual_machine(instance, [ 684.641174] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 684.641174] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] vif_infos = vmwarevif.get_vif_info(self._session, [ 684.641174] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 684.641478] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] for vif in network_info: [ 684.641478] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 684.641478] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] return self._sync_wrapper(fn, *args, **kwargs) [ 684.641478] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 684.641478] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] self.wait() [ 684.641478] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 684.641478] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] self[:] = self._gt.wait() [ 684.641478] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 684.641478] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] return self._exit_event.wait() [ 684.641478] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 684.641478] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] current.throw(*self._exc) [ 684.641478] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 684.641478] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] result = function(*args, **kwargs) [ 684.641877] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 684.641877] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] return func(*args, **kwargs) [ 684.641877] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 684.641877] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] raise e [ 684.641877] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 684.641877] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] nwinfo = self.network_api.allocate_for_instance( [ 684.641877] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 684.641877] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] created_port_ids = self._update_ports_for_instance( [ 684.641877] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 684.641877] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] with excutils.save_and_reraise_exception(): [ 684.641877] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 684.641877] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] self.force_reraise() [ 684.641877] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 684.642257] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] raise self.value [ 684.642257] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 684.642257] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] updated_port = self._update_port( [ 684.642257] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 684.642257] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] _ensure_no_port_binding_failure(port) [ 684.642257] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 684.642257] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] raise exception.PortBindingFailed(port_id=port['id']) [ 684.642257] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] nova.exception.PortBindingFailed: Binding failed for port 51e4aed6-7891-4427-90b4-0458a81ef471, please check neutron logs for more information. [ 684.642257] env[63088]: ERROR nova.compute.manager [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] [ 684.642257] env[63088]: DEBUG nova.compute.utils [None req-32cec7a8-18a7-43df-b82f-d466be0dcd4a tempest-AttachInterfacesUnderV243Test-1205880501 tempest-AttachInterfacesUnderV243Test-1205880501-project-member] [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] Binding failed for port 51e4aed6-7891-4427-90b4-0458a81ef471, please check neutron logs for more information. {{(pid=63088) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 684.643079] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bb8843aa-5c19-4289-a41c-17cbfcb47a00 tempest-ServerActionsTestOtherA-1614359098 tempest-ServerActionsTestOtherA-1614359098-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.686s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 684.646307] env[63088]: DEBUG nova.compute.manager [None req-32cec7a8-18a7-43df-b82f-d466be0dcd4a tempest-AttachInterfacesUnderV243Test-1205880501 tempest-AttachInterfacesUnderV243Test-1205880501-project-member] [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] Build of instance 6137d44e-9945-479d-9cf3-4a36bfa0e32c was re-scheduled: Binding failed for port 51e4aed6-7891-4427-90b4-0458a81ef471, please check neutron logs for more information. {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 684.648403] env[63088]: DEBUG nova.compute.manager [None req-32cec7a8-18a7-43df-b82f-d466be0dcd4a tempest-AttachInterfacesUnderV243Test-1205880501 tempest-AttachInterfacesUnderV243Test-1205880501-project-member] [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] Unplugging VIFs for instance {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 684.648403] env[63088]: DEBUG oslo_concurrency.lockutils [None req-32cec7a8-18a7-43df-b82f-d466be0dcd4a tempest-AttachInterfacesUnderV243Test-1205880501 tempest-AttachInterfacesUnderV243Test-1205880501-project-member] Acquiring lock "refresh_cache-6137d44e-9945-479d-9cf3-4a36bfa0e32c" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 684.648403] env[63088]: DEBUG oslo_concurrency.lockutils [None req-32cec7a8-18a7-43df-b82f-d466be0dcd4a tempest-AttachInterfacesUnderV243Test-1205880501 tempest-AttachInterfacesUnderV243Test-1205880501-project-member] Acquired lock "refresh_cache-6137d44e-9945-479d-9cf3-4a36bfa0e32c" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 684.648403] env[63088]: DEBUG nova.network.neutron [None req-32cec7a8-18a7-43df-b82f-d466be0dcd4a tempest-AttachInterfacesUnderV243Test-1205880501 tempest-AttachInterfacesUnderV243Test-1205880501-project-member] [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 684.872215] env[63088]: INFO nova.compute.manager [None req-442e0a41-8be2-46c5-9922-69b38e5ecc58 tempest-ServerAddressesNegativeTestJSON-62458283 tempest-ServerAddressesNegativeTestJSON-62458283-project-member] [instance: be5c7b15-c036-4a20-a21b-74bba67c9a2c] Took 1.02 seconds to deallocate network for instance. [ 684.934313] env[63088]: DEBUG nova.compute.manager [None req-0c6bc062-cb90-4d50-aabf-8c0503771f7e tempest-InstanceActionsNegativeTestJSON-1719444358 tempest-InstanceActionsNegativeTestJSON-1719444358-project-member] [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 685.177257] env[63088]: DEBUG nova.network.neutron [None req-32cec7a8-18a7-43df-b82f-d466be0dcd4a tempest-AttachInterfacesUnderV243Test-1205880501 tempest-AttachInterfacesUnderV243Test-1205880501-project-member] [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 685.266569] env[63088]: DEBUG nova.network.neutron [None req-32cec7a8-18a7-43df-b82f-d466be0dcd4a tempest-AttachInterfacesUnderV243Test-1205880501 tempest-AttachInterfacesUnderV243Test-1205880501-project-member] [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 685.452385] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0c6bc062-cb90-4d50-aabf-8c0503771f7e tempest-InstanceActionsNegativeTestJSON-1719444358 tempest-InstanceActionsNegativeTestJSON-1719444358-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 685.473053] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-681f7801-31da-4cbe-9903-56c1d61760a1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.480789] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c96ea110-fd86-4e26-afff-58e009abc752 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.512116] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fcb69e7-1914-43ad-9459-4b93d92d15f8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.519384] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-295e8e50-3ef9-4ca7-a3bb-1582f52d095f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.532400] env[63088]: DEBUG nova.compute.provider_tree [None req-bb8843aa-5c19-4289-a41c-17cbfcb47a00 tempest-ServerActionsTestOtherA-1614359098 tempest-ServerActionsTestOtherA-1614359098-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 685.769608] env[63088]: DEBUG oslo_concurrency.lockutils [None req-32cec7a8-18a7-43df-b82f-d466be0dcd4a tempest-AttachInterfacesUnderV243Test-1205880501 tempest-AttachInterfacesUnderV243Test-1205880501-project-member] Releasing lock "refresh_cache-6137d44e-9945-479d-9cf3-4a36bfa0e32c" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 685.769859] env[63088]: DEBUG nova.compute.manager [None req-32cec7a8-18a7-43df-b82f-d466be0dcd4a tempest-AttachInterfacesUnderV243Test-1205880501 tempest-AttachInterfacesUnderV243Test-1205880501-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 685.770057] env[63088]: DEBUG nova.compute.manager [None req-32cec7a8-18a7-43df-b82f-d466be0dcd4a tempest-AttachInterfacesUnderV243Test-1205880501 tempest-AttachInterfacesUnderV243Test-1205880501-project-member] [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 685.770236] env[63088]: DEBUG nova.network.neutron [None req-32cec7a8-18a7-43df-b82f-d466be0dcd4a tempest-AttachInterfacesUnderV243Test-1205880501 tempest-AttachInterfacesUnderV243Test-1205880501-project-member] [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 685.788979] env[63088]: DEBUG nova.network.neutron [None req-32cec7a8-18a7-43df-b82f-d466be0dcd4a tempest-AttachInterfacesUnderV243Test-1205880501 tempest-AttachInterfacesUnderV243Test-1205880501-project-member] [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 685.907822] env[63088]: INFO nova.scheduler.client.report [None req-442e0a41-8be2-46c5-9922-69b38e5ecc58 tempest-ServerAddressesNegativeTestJSON-62458283 tempest-ServerAddressesNegativeTestJSON-62458283-project-member] Deleted allocations for instance be5c7b15-c036-4a20-a21b-74bba67c9a2c [ 686.035267] env[63088]: DEBUG nova.scheduler.client.report [None req-bb8843aa-5c19-4289-a41c-17cbfcb47a00 tempest-ServerActionsTestOtherA-1614359098 tempest-ServerActionsTestOtherA-1614359098-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 686.291763] env[63088]: DEBUG nova.network.neutron [None req-32cec7a8-18a7-43df-b82f-d466be0dcd4a tempest-AttachInterfacesUnderV243Test-1205880501 tempest-AttachInterfacesUnderV243Test-1205880501-project-member] [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 686.417179] env[63088]: DEBUG oslo_concurrency.lockutils [None req-442e0a41-8be2-46c5-9922-69b38e5ecc58 tempest-ServerAddressesNegativeTestJSON-62458283 tempest-ServerAddressesNegativeTestJSON-62458283-project-member] Lock "be5c7b15-c036-4a20-a21b-74bba67c9a2c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 101.337s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 686.540575] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bb8843aa-5c19-4289-a41c-17cbfcb47a00 tempest-ServerActionsTestOtherA-1614359098 tempest-ServerActionsTestOtherA-1614359098-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.897s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 686.541262] env[63088]: ERROR nova.compute.manager [None req-bb8843aa-5c19-4289-a41c-17cbfcb47a00 tempest-ServerActionsTestOtherA-1614359098 tempest-ServerActionsTestOtherA-1614359098-project-member] [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d9f05fa0-ab0e-479e-ad2f-09a2ece1af2d, please check neutron logs for more information. [ 686.541262] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] Traceback (most recent call last): [ 686.541262] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 686.541262] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] self.driver.spawn(context, instance, image_meta, [ 686.541262] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 686.541262] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 686.541262] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 686.541262] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] vm_ref = self.build_virtual_machine(instance, [ 686.541262] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 686.541262] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] vif_infos = vmwarevif.get_vif_info(self._session, [ 686.541262] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 686.541642] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] for vif in network_info: [ 686.541642] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 686.541642] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] return self._sync_wrapper(fn, *args, **kwargs) [ 686.541642] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 686.541642] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] self.wait() [ 686.541642] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 686.541642] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] self[:] = self._gt.wait() [ 686.541642] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 686.541642] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] return self._exit_event.wait() [ 686.541642] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 686.541642] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] current.throw(*self._exc) [ 686.541642] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 686.541642] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] result = function(*args, **kwargs) [ 686.542046] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 686.542046] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] return func(*args, **kwargs) [ 686.542046] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 686.542046] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] raise e [ 686.542046] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 686.542046] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] nwinfo = self.network_api.allocate_for_instance( [ 686.542046] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 686.542046] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] created_port_ids = self._update_ports_for_instance( [ 686.542046] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 686.542046] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] with excutils.save_and_reraise_exception(): [ 686.542046] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 686.542046] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] self.force_reraise() [ 686.542046] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 686.542400] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] raise self.value [ 686.542400] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 686.542400] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] updated_port = self._update_port( [ 686.542400] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 686.542400] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] _ensure_no_port_binding_failure(port) [ 686.542400] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 686.542400] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] raise exception.PortBindingFailed(port_id=port['id']) [ 686.542400] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] nova.exception.PortBindingFailed: Binding failed for port d9f05fa0-ab0e-479e-ad2f-09a2ece1af2d, please check neutron logs for more information. [ 686.542400] env[63088]: ERROR nova.compute.manager [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] [ 686.542400] env[63088]: DEBUG nova.compute.utils [None req-bb8843aa-5c19-4289-a41c-17cbfcb47a00 tempest-ServerActionsTestOtherA-1614359098 tempest-ServerActionsTestOtherA-1614359098-project-member] [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] Binding failed for port d9f05fa0-ab0e-479e-ad2f-09a2ece1af2d, please check neutron logs for more information. {{(pid=63088) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 686.543183] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 16.382s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 686.544862] env[63088]: DEBUG nova.compute.manager [None req-bb8843aa-5c19-4289-a41c-17cbfcb47a00 tempest-ServerActionsTestOtherA-1614359098 tempest-ServerActionsTestOtherA-1614359098-project-member] [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] Build of instance 90c1832a-47c5-4f09-9921-21e15cb6fbb6 was re-scheduled: Binding failed for port d9f05fa0-ab0e-479e-ad2f-09a2ece1af2d, please check neutron logs for more information. {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 686.545366] env[63088]: DEBUG nova.compute.manager [None req-bb8843aa-5c19-4289-a41c-17cbfcb47a00 tempest-ServerActionsTestOtherA-1614359098 tempest-ServerActionsTestOtherA-1614359098-project-member] [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] Unplugging VIFs for instance {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 686.545596] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bb8843aa-5c19-4289-a41c-17cbfcb47a00 tempest-ServerActionsTestOtherA-1614359098 tempest-ServerActionsTestOtherA-1614359098-project-member] Acquiring lock "refresh_cache-90c1832a-47c5-4f09-9921-21e15cb6fbb6" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 686.545739] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bb8843aa-5c19-4289-a41c-17cbfcb47a00 tempest-ServerActionsTestOtherA-1614359098 tempest-ServerActionsTestOtherA-1614359098-project-member] Acquired lock "refresh_cache-90c1832a-47c5-4f09-9921-21e15cb6fbb6" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 686.545904] env[63088]: DEBUG nova.network.neutron [None req-bb8843aa-5c19-4289-a41c-17cbfcb47a00 tempest-ServerActionsTestOtherA-1614359098 tempest-ServerActionsTestOtherA-1614359098-project-member] [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 686.795489] env[63088]: INFO nova.compute.manager [None req-32cec7a8-18a7-43df-b82f-d466be0dcd4a tempest-AttachInterfacesUnderV243Test-1205880501 tempest-AttachInterfacesUnderV243Test-1205880501-project-member] [instance: 6137d44e-9945-479d-9cf3-4a36bfa0e32c] Took 1.02 seconds to deallocate network for instance. [ 686.919314] env[63088]: DEBUG nova.compute.manager [None req-a2a0ac90-0e7a-45d1-a070-f7ee6be01d68 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 687.065332] env[63088]: DEBUG nova.network.neutron [None req-bb8843aa-5c19-4289-a41c-17cbfcb47a00 tempest-ServerActionsTestOtherA-1614359098 tempest-ServerActionsTestOtherA-1614359098-project-member] [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 687.138269] env[63088]: DEBUG nova.network.neutron [None req-bb8843aa-5c19-4289-a41c-17cbfcb47a00 tempest-ServerActionsTestOtherA-1614359098 tempest-ServerActionsTestOtherA-1614359098-project-member] [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 687.445265] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a2a0ac90-0e7a-45d1-a070-f7ee6be01d68 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 687.641929] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bb8843aa-5c19-4289-a41c-17cbfcb47a00 tempest-ServerActionsTestOtherA-1614359098 tempest-ServerActionsTestOtherA-1614359098-project-member] Releasing lock "refresh_cache-90c1832a-47c5-4f09-9921-21e15cb6fbb6" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 687.642210] env[63088]: DEBUG nova.compute.manager [None req-bb8843aa-5c19-4289-a41c-17cbfcb47a00 tempest-ServerActionsTestOtherA-1614359098 tempest-ServerActionsTestOtherA-1614359098-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 687.642396] env[63088]: DEBUG nova.compute.manager [None req-bb8843aa-5c19-4289-a41c-17cbfcb47a00 tempest-ServerActionsTestOtherA-1614359098 tempest-ServerActionsTestOtherA-1614359098-project-member] [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 687.642561] env[63088]: DEBUG nova.network.neutron [None req-bb8843aa-5c19-4289-a41c-17cbfcb47a00 tempest-ServerActionsTestOtherA-1614359098 tempest-ServerActionsTestOtherA-1614359098-project-member] [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 687.670961] env[63088]: DEBUG nova.network.neutron [None req-bb8843aa-5c19-4289-a41c-17cbfcb47a00 tempest-ServerActionsTestOtherA-1614359098 tempest-ServerActionsTestOtherA-1614359098-project-member] [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 687.834017] env[63088]: INFO nova.scheduler.client.report [None req-32cec7a8-18a7-43df-b82f-d466be0dcd4a tempest-AttachInterfacesUnderV243Test-1205880501 tempest-AttachInterfacesUnderV243Test-1205880501-project-member] Deleted allocations for instance 6137d44e-9945-479d-9cf3-4a36bfa0e32c [ 688.075620] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance 6137d44e-9945-479d-9cf3-4a36bfa0e32c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 688.173219] env[63088]: DEBUG nova.network.neutron [None req-bb8843aa-5c19-4289-a41c-17cbfcb47a00 tempest-ServerActionsTestOtherA-1614359098 tempest-ServerActionsTestOtherA-1614359098-project-member] [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 688.344064] env[63088]: DEBUG oslo_concurrency.lockutils [None req-32cec7a8-18a7-43df-b82f-d466be0dcd4a tempest-AttachInterfacesUnderV243Test-1205880501 tempest-AttachInterfacesUnderV243Test-1205880501-project-member] Lock "6137d44e-9945-479d-9cf3-4a36bfa0e32c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 97.253s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 688.577802] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance 90c1832a-47c5-4f09-9921-21e15cb6fbb6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 688.578123] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance 0f515940-fa10-4eb8-970d-ead4b996021c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 688.578123] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance d9a4d198-4aee-4882-bee8-f6cde14745a2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 688.578243] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance 6eb3e767-ff7d-4760-a036-a55647afac0f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 688.675913] env[63088]: INFO nova.compute.manager [None req-bb8843aa-5c19-4289-a41c-17cbfcb47a00 tempest-ServerActionsTestOtherA-1614359098 tempest-ServerActionsTestOtherA-1614359098-project-member] [instance: 90c1832a-47c5-4f09-9921-21e15cb6fbb6] Took 1.03 seconds to deallocate network for instance. [ 688.847183] env[63088]: DEBUG nova.compute.manager [None req-ae448e13-3c80-41d5-9b1f-e650107512d4 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 689.081498] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance 02da88f0-7e3a-4eea-a88d-59a990c69f82 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 689.371414] env[63088]: DEBUG oslo_concurrency.lockutils [None req-ae448e13-3c80-41d5-9b1f-e650107512d4 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 689.585929] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance c67b276f-07db-4cde-9f1c-efccc4d11c7d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 689.710997] env[63088]: INFO nova.scheduler.client.report [None req-bb8843aa-5c19-4289-a41c-17cbfcb47a00 tempest-ServerActionsTestOtherA-1614359098 tempest-ServerActionsTestOtherA-1614359098-project-member] Deleted allocations for instance 90c1832a-47c5-4f09-9921-21e15cb6fbb6 [ 690.090122] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance 1219fe86-f218-4a2b-84ba-4a377829386d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 690.220128] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bb8843aa-5c19-4289-a41c-17cbfcb47a00 tempest-ServerActionsTestOtherA-1614359098 tempest-ServerActionsTestOtherA-1614359098-project-member] Lock "90c1832a-47c5-4f09-9921-21e15cb6fbb6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 98.474s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 690.593374] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance 20611ed4-97b3-4d6d-bef7-c13e1235c401 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 690.722571] env[63088]: DEBUG nova.compute.manager [None req-7af44f64-0067-49a1-ac0c-3ee5b8047867 tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 691.097280] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance eb42c667-b0ae-4807-8f69-88dd671aa7d7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 691.253059] env[63088]: DEBUG oslo_concurrency.lockutils [None req-7af44f64-0067-49a1-ac0c-3ee5b8047867 tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 691.601759] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance 2b43620d-94ac-4c56-87df-4066f5d24088 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 692.111610] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance f2b4485c-ebcb-4063-a25a-e2efe6d1758f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 692.615225] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance c63c1740-fcec-47fa-a7b9-986b83c4abd9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 693.119990] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance 407db84f-3322-42bc-baac-5528e856af8d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 693.624831] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 693.723872] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b5c531a2-a229-4998-9a9b-9ebb1e7afff2 tempest-ServerGroupTestJSON-1534054659 tempest-ServerGroupTestJSON-1534054659-project-member] Acquiring lock "53ea5c01-4be2-44a8-b74f-0d74031dfdd7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 693.724429] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b5c531a2-a229-4998-9a9b-9ebb1e7afff2 tempest-ServerGroupTestJSON-1534054659 tempest-ServerGroupTestJSON-1534054659-project-member] Lock "53ea5c01-4be2-44a8-b74f-0d74031dfdd7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 694.128086] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance 641eed47-54f6-4c08-a8fd-f06bc5f7fb56 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 694.629995] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance b82c89ad-f7d6-4c04-a48e-acdab29ea5e7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 695.132845] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 695.337547] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Acquiring lock "db032da3-d334-486f-a428-af1c8a3c360a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 695.337784] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Lock "db032da3-d334-486f-a428-af1c8a3c360a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 695.635869] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance 17725a7b-ccd3-477e-a7a6-b96af45d6d1f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 696.140115] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance 7b5cc7b5-98b9-4344-b714-270dca798799 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 696.643375] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance d851a26b-48aa-4ffd-9c23-cea27543e0fa has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 697.148373] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance 43ded929-6890-4296-ae90-230f7c8b2595 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 697.651112] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 698.154184] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance 87361d1c-e9cd-4177-bf3d-5d7551192073 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 698.657200] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance e6a72b08-38b6-44ae-8911-6b39d50ee1c3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 698.657534] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Total usable vcpus: 48, total allocated vcpus: 3 {{(pid=63088) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 698.657589] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1088MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=3 pci_stats=[] {{(pid=63088) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 698.928281] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecb8afd5-1883-487c-8168-2c00a0d09ce6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.936999] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d98ddb8f-ddf3-4912-b9b0-b0883bb4fc32 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.966663] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d4e64de-2cec-448a-bdac-2bf64fd09641 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.974158] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b1ad534-915f-4273-86c8-22114fa4b37f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.987897] env[63088]: DEBUG nova.compute.provider_tree [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 699.491297] env[63088]: DEBUG nova.scheduler.client.report [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 699.995807] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63088) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 699.996079] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 13.453s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 699.996338] env[63088]: DEBUG oslo_concurrency.lockutils [None req-18116a16-f5bf-4db6-aaca-b64d4351d5d4 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 27.633s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 700.776859] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98e77160-c12e-45fc-86e3-697c2ba5eed8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.784494] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11b15721-3aa2-4278-ac66-58197c97b55f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.813935] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d002da6-6178-421d-b7ce-e59a222be99c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.821372] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78afeb6e-e34a-4f65-bb4c-76c73fdae0c3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.834585] env[63088]: DEBUG nova.compute.provider_tree [None req-18116a16-f5bf-4db6-aaca-b64d4351d5d4 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 701.338087] env[63088]: DEBUG nova.scheduler.client.report [None req-18116a16-f5bf-4db6-aaca-b64d4351d5d4 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 701.842636] env[63088]: DEBUG oslo_concurrency.lockutils [None req-18116a16-f5bf-4db6-aaca-b64d4351d5d4 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.846s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 701.843302] env[63088]: ERROR nova.compute.manager [None req-18116a16-f5bf-4db6-aaca-b64d4351d5d4 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 29c388e4-0159-4c3d-8f7d-9a0cd5ae3d04, please check neutron logs for more information. [ 701.843302] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] Traceback (most recent call last): [ 701.843302] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 701.843302] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] self.driver.spawn(context, instance, image_meta, [ 701.843302] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 701.843302] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 701.843302] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 701.843302] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] vm_ref = self.build_virtual_machine(instance, [ 701.843302] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 701.843302] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] vif_infos = vmwarevif.get_vif_info(self._session, [ 701.843302] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 701.843654] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] for vif in network_info: [ 701.843654] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 701.843654] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] return self._sync_wrapper(fn, *args, **kwargs) [ 701.843654] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 701.843654] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] self.wait() [ 701.843654] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 701.843654] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] self[:] = self._gt.wait() [ 701.843654] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 701.843654] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] return self._exit_event.wait() [ 701.843654] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 701.843654] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] current.throw(*self._exc) [ 701.843654] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 701.843654] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] result = function(*args, **kwargs) [ 701.844050] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 701.844050] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] return func(*args, **kwargs) [ 701.844050] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 701.844050] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] raise e [ 701.844050] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 701.844050] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] nwinfo = self.network_api.allocate_for_instance( [ 701.844050] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 701.844050] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] created_port_ids = self._update_ports_for_instance( [ 701.844050] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 701.844050] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] with excutils.save_and_reraise_exception(): [ 701.844050] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 701.844050] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] self.force_reraise() [ 701.844050] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 701.844374] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] raise self.value [ 701.844374] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 701.844374] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] updated_port = self._update_port( [ 701.844374] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 701.844374] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] _ensure_no_port_binding_failure(port) [ 701.844374] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 701.844374] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] raise exception.PortBindingFailed(port_id=port['id']) [ 701.844374] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] nova.exception.PortBindingFailed: Binding failed for port 29c388e4-0159-4c3d-8f7d-9a0cd5ae3d04, please check neutron logs for more information. [ 701.844374] env[63088]: ERROR nova.compute.manager [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] [ 701.844374] env[63088]: DEBUG nova.compute.utils [None req-18116a16-f5bf-4db6-aaca-b64d4351d5d4 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] Binding failed for port 29c388e4-0159-4c3d-8f7d-9a0cd5ae3d04, please check neutron logs for more information. {{(pid=63088) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 701.845277] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0b29042d-57c3-47a9-a926-f67d685a4740 tempest-TenantUsagesTestJSON-1054626739 tempest-TenantUsagesTestJSON-1054626739-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.050s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 701.846657] env[63088]: INFO nova.compute.claims [None req-0b29042d-57c3-47a9-a926-f67d685a4740 tempest-TenantUsagesTestJSON-1054626739 tempest-TenantUsagesTestJSON-1054626739-project-member] [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 701.851068] env[63088]: DEBUG nova.compute.manager [None req-18116a16-f5bf-4db6-aaca-b64d4351d5d4 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] Build of instance 0f515940-fa10-4eb8-970d-ead4b996021c was re-scheduled: Binding failed for port 29c388e4-0159-4c3d-8f7d-9a0cd5ae3d04, please check neutron logs for more information. {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 701.851068] env[63088]: DEBUG nova.compute.manager [None req-18116a16-f5bf-4db6-aaca-b64d4351d5d4 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] Unplugging VIFs for instance {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 701.851335] env[63088]: DEBUG oslo_concurrency.lockutils [None req-18116a16-f5bf-4db6-aaca-b64d4351d5d4 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Acquiring lock "refresh_cache-0f515940-fa10-4eb8-970d-ead4b996021c" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 701.851406] env[63088]: DEBUG oslo_concurrency.lockutils [None req-18116a16-f5bf-4db6-aaca-b64d4351d5d4 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Acquired lock "refresh_cache-0f515940-fa10-4eb8-970d-ead4b996021c" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 701.851568] env[63088]: DEBUG nova.network.neutron [None req-18116a16-f5bf-4db6-aaca-b64d4351d5d4 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 702.370689] env[63088]: DEBUG nova.network.neutron [None req-18116a16-f5bf-4db6-aaca-b64d4351d5d4 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 702.467808] env[63088]: DEBUG nova.network.neutron [None req-18116a16-f5bf-4db6-aaca-b64d4351d5d4 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 702.970096] env[63088]: DEBUG oslo_concurrency.lockutils [None req-18116a16-f5bf-4db6-aaca-b64d4351d5d4 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Releasing lock "refresh_cache-0f515940-fa10-4eb8-970d-ead4b996021c" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 702.970415] env[63088]: DEBUG nova.compute.manager [None req-18116a16-f5bf-4db6-aaca-b64d4351d5d4 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 702.970540] env[63088]: DEBUG nova.compute.manager [None req-18116a16-f5bf-4db6-aaca-b64d4351d5d4 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 702.970654] env[63088]: DEBUG nova.network.neutron [None req-18116a16-f5bf-4db6-aaca-b64d4351d5d4 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 702.987941] env[63088]: DEBUG nova.network.neutron [None req-18116a16-f5bf-4db6-aaca-b64d4351d5d4 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 703.135990] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f6bafd5-3de2-4bcf-9405-0bcbbdaa99f6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.145094] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ac5bd58-d04d-4553-a1f3-d8bd26a3c621 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.172789] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9267147b-c074-4ac7-ad71-36f25c67e8dd {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.180530] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0116a6a6-30c6-475e-90b0-466035e02152 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.193530] env[63088]: DEBUG nova.compute.provider_tree [None req-0b29042d-57c3-47a9-a926-f67d685a4740 tempest-TenantUsagesTestJSON-1054626739 tempest-TenantUsagesTestJSON-1054626739-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 703.490289] env[63088]: DEBUG nova.network.neutron [None req-18116a16-f5bf-4db6-aaca-b64d4351d5d4 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 703.696837] env[63088]: DEBUG nova.scheduler.client.report [None req-0b29042d-57c3-47a9-a926-f67d685a4740 tempest-TenantUsagesTestJSON-1054626739 tempest-TenantUsagesTestJSON-1054626739-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 703.993441] env[63088]: INFO nova.compute.manager [None req-18116a16-f5bf-4db6-aaca-b64d4351d5d4 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: 0f515940-fa10-4eb8-970d-ead4b996021c] Took 1.02 seconds to deallocate network for instance. [ 704.201998] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0b29042d-57c3-47a9-a926-f67d685a4740 tempest-TenantUsagesTestJSON-1054626739 tempest-TenantUsagesTestJSON-1054626739-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.357s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 704.202556] env[63088]: DEBUG nova.compute.manager [None req-0b29042d-57c3-47a9-a926-f67d685a4740 tempest-TenantUsagesTestJSON-1054626739 tempest-TenantUsagesTestJSON-1054626739-project-member] [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 704.207024] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f237740d-6eaf-4423-b770-dd9775fe6fd5 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 26.529s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 704.707232] env[63088]: DEBUG nova.compute.utils [None req-0b29042d-57c3-47a9-a926-f67d685a4740 tempest-TenantUsagesTestJSON-1054626739 tempest-TenantUsagesTestJSON-1054626739-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 704.708981] env[63088]: DEBUG nova.compute.manager [None req-0b29042d-57c3-47a9-a926-f67d685a4740 tempest-TenantUsagesTestJSON-1054626739 tempest-TenantUsagesTestJSON-1054626739-project-member] [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 704.709216] env[63088]: DEBUG nova.network.neutron [None req-0b29042d-57c3-47a9-a926-f67d685a4740 tempest-TenantUsagesTestJSON-1054626739 tempest-TenantUsagesTestJSON-1054626739-project-member] [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 704.778252] env[63088]: DEBUG nova.policy [None req-0b29042d-57c3-47a9-a926-f67d685a4740 tempest-TenantUsagesTestJSON-1054626739 tempest-TenantUsagesTestJSON-1054626739-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1d2e272b79c8445d9528ba13ac9063b8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3e8d3adcaba7409aa3c449991d5f1ed2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 705.003150] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3eea97ae-892b-457c-bedf-0a4b8f65e2b4 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.014873] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3c57633-412b-4535-82a1-54a1c3a3e249 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.044691] env[63088]: INFO nova.scheduler.client.report [None req-18116a16-f5bf-4db6-aaca-b64d4351d5d4 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Deleted allocations for instance 0f515940-fa10-4eb8-970d-ead4b996021c [ 705.050541] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d29c278-30f1-4486-b388-83a2dd60c4f1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.058981] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6137321b-bb0c-48d7-9415-a81fb153f48a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.073930] env[63088]: DEBUG nova.compute.provider_tree [None req-f237740d-6eaf-4423-b770-dd9775fe6fd5 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 705.137443] env[63088]: DEBUG nova.network.neutron [None req-0b29042d-57c3-47a9-a926-f67d685a4740 tempest-TenantUsagesTestJSON-1054626739 tempest-TenantUsagesTestJSON-1054626739-project-member] [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] Successfully created port: 32cece70-208a-4d57-bcdf-d63d938be4d3 {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 705.218426] env[63088]: DEBUG nova.compute.manager [None req-0b29042d-57c3-47a9-a926-f67d685a4740 tempest-TenantUsagesTestJSON-1054626739 tempest-TenantUsagesTestJSON-1054626739-project-member] [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 705.555855] env[63088]: DEBUG oslo_concurrency.lockutils [None req-18116a16-f5bf-4db6-aaca-b64d4351d5d4 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Lock "0f515940-fa10-4eb8-970d-ead4b996021c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 112.941s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 705.579264] env[63088]: DEBUG nova.scheduler.client.report [None req-f237740d-6eaf-4423-b770-dd9775fe6fd5 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 705.919249] env[63088]: DEBUG nova.compute.manager [req-240751f2-478d-4dba-a228-278dc60b1ffe req-b320b21b-37eb-4cd4-82b3-a99f537f24a7 service nova] [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] Received event network-changed-32cece70-208a-4d57-bcdf-d63d938be4d3 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 705.919249] env[63088]: DEBUG nova.compute.manager [req-240751f2-478d-4dba-a228-278dc60b1ffe req-b320b21b-37eb-4cd4-82b3-a99f537f24a7 service nova] [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] Refreshing instance network info cache due to event network-changed-32cece70-208a-4d57-bcdf-d63d938be4d3. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 705.919249] env[63088]: DEBUG oslo_concurrency.lockutils [req-240751f2-478d-4dba-a228-278dc60b1ffe req-b320b21b-37eb-4cd4-82b3-a99f537f24a7 service nova] Acquiring lock "refresh_cache-02da88f0-7e3a-4eea-a88d-59a990c69f82" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 705.919249] env[63088]: DEBUG oslo_concurrency.lockutils [req-240751f2-478d-4dba-a228-278dc60b1ffe req-b320b21b-37eb-4cd4-82b3-a99f537f24a7 service nova] Acquired lock "refresh_cache-02da88f0-7e3a-4eea-a88d-59a990c69f82" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 705.919249] env[63088]: DEBUG nova.network.neutron [req-240751f2-478d-4dba-a228-278dc60b1ffe req-b320b21b-37eb-4cd4-82b3-a99f537f24a7 service nova] [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] Refreshing network info cache for port 32cece70-208a-4d57-bcdf-d63d938be4d3 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 706.060706] env[63088]: DEBUG nova.compute.manager [None req-03263d3e-caae-4f24-95aa-ac7aedd64259 tempest-ImagesOneServerTestJSON-44065314 tempest-ImagesOneServerTestJSON-44065314-project-member] [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 706.081705] env[63088]: ERROR nova.compute.manager [None req-0b29042d-57c3-47a9-a926-f67d685a4740 tempest-TenantUsagesTestJSON-1054626739 tempest-TenantUsagesTestJSON-1054626739-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 32cece70-208a-4d57-bcdf-d63d938be4d3, please check neutron logs for more information. [ 706.081705] env[63088]: ERROR nova.compute.manager Traceback (most recent call last): [ 706.081705] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 706.081705] env[63088]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 706.081705] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 706.081705] env[63088]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 706.081705] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 706.081705] env[63088]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 706.081705] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 706.081705] env[63088]: ERROR nova.compute.manager self.force_reraise() [ 706.081705] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 706.081705] env[63088]: ERROR nova.compute.manager raise self.value [ 706.081705] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 706.081705] env[63088]: ERROR nova.compute.manager updated_port = self._update_port( [ 706.081705] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 706.081705] env[63088]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 706.082222] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 706.082222] env[63088]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 706.082222] env[63088]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 32cece70-208a-4d57-bcdf-d63d938be4d3, please check neutron logs for more information. [ 706.082222] env[63088]: ERROR nova.compute.manager [ 706.082222] env[63088]: Traceback (most recent call last): [ 706.082222] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 706.082222] env[63088]: listener.cb(fileno) [ 706.082222] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 706.082222] env[63088]: result = function(*args, **kwargs) [ 706.082478] env[63088]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 706.082478] env[63088]: return func(*args, **kwargs) [ 706.082478] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 706.082478] env[63088]: raise e [ 706.082478] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 706.082478] env[63088]: nwinfo = self.network_api.allocate_for_instance( [ 706.082478] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 706.082478] env[63088]: created_port_ids = self._update_ports_for_instance( [ 706.082478] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 706.082478] env[63088]: with excutils.save_and_reraise_exception(): [ 706.082478] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 706.082478] env[63088]: self.force_reraise() [ 706.082478] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 706.082478] env[63088]: raise self.value [ 706.082478] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 706.082478] env[63088]: updated_port = self._update_port( [ 706.082478] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 706.082478] env[63088]: _ensure_no_port_binding_failure(port) [ 706.082478] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 706.082478] env[63088]: raise exception.PortBindingFailed(port_id=port['id']) [ 706.082478] env[63088]: nova.exception.PortBindingFailed: Binding failed for port 32cece70-208a-4d57-bcdf-d63d938be4d3, please check neutron logs for more information. [ 706.082478] env[63088]: Removing descriptor: 17 [ 706.084212] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f237740d-6eaf-4423-b770-dd9775fe6fd5 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.878s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 706.084798] env[63088]: ERROR nova.compute.manager [None req-f237740d-6eaf-4423-b770-dd9775fe6fd5 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e454d679-a450-4e9d-81d6-228f3fd67f14, please check neutron logs for more information. [ 706.084798] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] Traceback (most recent call last): [ 706.084798] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 706.084798] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] self.driver.spawn(context, instance, image_meta, [ 706.084798] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 706.084798] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 706.084798] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 706.084798] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] vm_ref = self.build_virtual_machine(instance, [ 706.084798] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 706.084798] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] vif_infos = vmwarevif.get_vif_info(self._session, [ 706.084798] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 706.085193] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] for vif in network_info: [ 706.085193] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 706.085193] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] return self._sync_wrapper(fn, *args, **kwargs) [ 706.085193] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 706.085193] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] self.wait() [ 706.085193] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 706.085193] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] self[:] = self._gt.wait() [ 706.085193] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 706.085193] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] return self._exit_event.wait() [ 706.085193] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 706.085193] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] current.throw(*self._exc) [ 706.085193] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 706.085193] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] result = function(*args, **kwargs) [ 706.085572] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 706.085572] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] return func(*args, **kwargs) [ 706.085572] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 706.085572] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] raise e [ 706.085572] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 706.085572] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] nwinfo = self.network_api.allocate_for_instance( [ 706.085572] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 706.085572] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] created_port_ids = self._update_ports_for_instance( [ 706.085572] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 706.085572] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] with excutils.save_and_reraise_exception(): [ 706.085572] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 706.085572] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] self.force_reraise() [ 706.085572] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 706.085940] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] raise self.value [ 706.085940] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 706.085940] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] updated_port = self._update_port( [ 706.085940] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 706.085940] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] _ensure_no_port_binding_failure(port) [ 706.085940] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 706.085940] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] raise exception.PortBindingFailed(port_id=port['id']) [ 706.085940] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] nova.exception.PortBindingFailed: Binding failed for port e454d679-a450-4e9d-81d6-228f3fd67f14, please check neutron logs for more information. [ 706.085940] env[63088]: ERROR nova.compute.manager [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] [ 706.085940] env[63088]: DEBUG nova.compute.utils [None req-f237740d-6eaf-4423-b770-dd9775fe6fd5 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] Binding failed for port e454d679-a450-4e9d-81d6-228f3fd67f14, please check neutron logs for more information. {{(pid=63088) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 706.086680] env[63088]: DEBUG oslo_concurrency.lockutils [None req-393edcfb-0595-4190-8ce0-4c14157dbbff tempest-ServerDiagnosticsNegativeTest-1508062492 tempest-ServerDiagnosticsNegativeTest-1508062492-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 26.843s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 706.089423] env[63088]: DEBUG nova.compute.manager [None req-f237740d-6eaf-4423-b770-dd9775fe6fd5 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] Build of instance d9a4d198-4aee-4882-bee8-f6cde14745a2 was re-scheduled: Binding failed for port e454d679-a450-4e9d-81d6-228f3fd67f14, please check neutron logs for more information. {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 706.089818] env[63088]: DEBUG nova.compute.manager [None req-f237740d-6eaf-4423-b770-dd9775fe6fd5 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] Unplugging VIFs for instance {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 706.090053] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f237740d-6eaf-4423-b770-dd9775fe6fd5 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Acquiring lock "refresh_cache-d9a4d198-4aee-4882-bee8-f6cde14745a2" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 706.090206] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f237740d-6eaf-4423-b770-dd9775fe6fd5 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Acquired lock "refresh_cache-d9a4d198-4aee-4882-bee8-f6cde14745a2" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 706.090362] env[63088]: DEBUG nova.network.neutron [None req-f237740d-6eaf-4423-b770-dd9775fe6fd5 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 706.234676] env[63088]: DEBUG nova.compute.manager [None req-0b29042d-57c3-47a9-a926-f67d685a4740 tempest-TenantUsagesTestJSON-1054626739 tempest-TenantUsagesTestJSON-1054626739-project-member] [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 706.260955] env[63088]: DEBUG nova.virt.hardware [None req-0b29042d-57c3-47a9-a926-f67d685a4740 tempest-TenantUsagesTestJSON-1054626739 tempest-TenantUsagesTestJSON-1054626739-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 706.261215] env[63088]: DEBUG nova.virt.hardware [None req-0b29042d-57c3-47a9-a926-f67d685a4740 tempest-TenantUsagesTestJSON-1054626739 tempest-TenantUsagesTestJSON-1054626739-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 706.261371] env[63088]: DEBUG nova.virt.hardware [None req-0b29042d-57c3-47a9-a926-f67d685a4740 tempest-TenantUsagesTestJSON-1054626739 tempest-TenantUsagesTestJSON-1054626739-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 706.261552] env[63088]: DEBUG nova.virt.hardware [None req-0b29042d-57c3-47a9-a926-f67d685a4740 tempest-TenantUsagesTestJSON-1054626739 tempest-TenantUsagesTestJSON-1054626739-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 706.261697] env[63088]: DEBUG nova.virt.hardware [None req-0b29042d-57c3-47a9-a926-f67d685a4740 tempest-TenantUsagesTestJSON-1054626739 tempest-TenantUsagesTestJSON-1054626739-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 706.261842] env[63088]: DEBUG nova.virt.hardware [None req-0b29042d-57c3-47a9-a926-f67d685a4740 tempest-TenantUsagesTestJSON-1054626739 tempest-TenantUsagesTestJSON-1054626739-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 706.262059] env[63088]: DEBUG nova.virt.hardware [None req-0b29042d-57c3-47a9-a926-f67d685a4740 tempest-TenantUsagesTestJSON-1054626739 tempest-TenantUsagesTestJSON-1054626739-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 706.262224] env[63088]: DEBUG nova.virt.hardware [None req-0b29042d-57c3-47a9-a926-f67d685a4740 tempest-TenantUsagesTestJSON-1054626739 tempest-TenantUsagesTestJSON-1054626739-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 706.262392] env[63088]: DEBUG nova.virt.hardware [None req-0b29042d-57c3-47a9-a926-f67d685a4740 tempest-TenantUsagesTestJSON-1054626739 tempest-TenantUsagesTestJSON-1054626739-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 706.262552] env[63088]: DEBUG nova.virt.hardware [None req-0b29042d-57c3-47a9-a926-f67d685a4740 tempest-TenantUsagesTestJSON-1054626739 tempest-TenantUsagesTestJSON-1054626739-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 706.262722] env[63088]: DEBUG nova.virt.hardware [None req-0b29042d-57c3-47a9-a926-f67d685a4740 tempest-TenantUsagesTestJSON-1054626739 tempest-TenantUsagesTestJSON-1054626739-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 706.263578] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8170d5d2-f883-4f03-90ad-fa18280a502c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.271427] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-920dd702-72fb-482e-a50f-8bb911c055dc {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.284956] env[63088]: ERROR nova.compute.manager [None req-0b29042d-57c3-47a9-a926-f67d685a4740 tempest-TenantUsagesTestJSON-1054626739 tempest-TenantUsagesTestJSON-1054626739-project-member] [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 32cece70-208a-4d57-bcdf-d63d938be4d3, please check neutron logs for more information. [ 706.284956] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] Traceback (most recent call last): [ 706.284956] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 706.284956] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] yield resources [ 706.284956] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 706.284956] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] self.driver.spawn(context, instance, image_meta, [ 706.284956] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 706.284956] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] self._vmops.spawn(context, instance, image_meta, injected_files, [ 706.284956] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 706.284956] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] vm_ref = self.build_virtual_machine(instance, [ 706.284956] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 706.285383] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] vif_infos = vmwarevif.get_vif_info(self._session, [ 706.285383] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 706.285383] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] for vif in network_info: [ 706.285383] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 706.285383] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] return self._sync_wrapper(fn, *args, **kwargs) [ 706.285383] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 706.285383] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] self.wait() [ 706.285383] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 706.285383] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] self[:] = self._gt.wait() [ 706.285383] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 706.285383] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] return self._exit_event.wait() [ 706.285383] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 706.285383] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] current.throw(*self._exc) [ 706.285749] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 706.285749] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] result = function(*args, **kwargs) [ 706.285749] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 706.285749] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] return func(*args, **kwargs) [ 706.285749] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 706.285749] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] raise e [ 706.285749] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 706.285749] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] nwinfo = self.network_api.allocate_for_instance( [ 706.285749] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 706.285749] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] created_port_ids = self._update_ports_for_instance( [ 706.285749] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 706.285749] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] with excutils.save_and_reraise_exception(): [ 706.285749] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 706.286143] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] self.force_reraise() [ 706.286143] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 706.286143] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] raise self.value [ 706.286143] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 706.286143] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] updated_port = self._update_port( [ 706.286143] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 706.286143] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] _ensure_no_port_binding_failure(port) [ 706.286143] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 706.286143] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] raise exception.PortBindingFailed(port_id=port['id']) [ 706.286143] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] nova.exception.PortBindingFailed: Binding failed for port 32cece70-208a-4d57-bcdf-d63d938be4d3, please check neutron logs for more information. [ 706.286143] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] [ 706.286143] env[63088]: INFO nova.compute.manager [None req-0b29042d-57c3-47a9-a926-f67d685a4740 tempest-TenantUsagesTestJSON-1054626739 tempest-TenantUsagesTestJSON-1054626739-project-member] [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] Terminating instance [ 706.287510] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0b29042d-57c3-47a9-a926-f67d685a4740 tempest-TenantUsagesTestJSON-1054626739 tempest-TenantUsagesTestJSON-1054626739-project-member] Acquiring lock "refresh_cache-02da88f0-7e3a-4eea-a88d-59a990c69f82" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 706.435709] env[63088]: DEBUG nova.network.neutron [req-240751f2-478d-4dba-a228-278dc60b1ffe req-b320b21b-37eb-4cd4-82b3-a99f537f24a7 service nova] [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 706.479659] env[63088]: DEBUG nova.network.neutron [req-240751f2-478d-4dba-a228-278dc60b1ffe req-b320b21b-37eb-4cd4-82b3-a99f537f24a7 service nova] [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 706.582298] env[63088]: DEBUG oslo_concurrency.lockutils [None req-03263d3e-caae-4f24-95aa-ac7aedd64259 tempest-ImagesOneServerTestJSON-44065314 tempest-ImagesOneServerTestJSON-44065314-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 706.611944] env[63088]: DEBUG nova.network.neutron [None req-f237740d-6eaf-4423-b770-dd9775fe6fd5 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 706.702318] env[63088]: DEBUG nova.network.neutron [None req-f237740d-6eaf-4423-b770-dd9775fe6fd5 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 706.894017] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c058a219-4d89-473d-9c6d-6666f19d3b3b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.901622] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e73d92b3-8681-4d1d-8fce-3a181d9a0924 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.931875] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30aef8b6-1b4e-4598-83de-16d6a6298f4d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.939499] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b88f5e48-7369-49d7-83c7-fba35f4222a4 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.953110] env[63088]: DEBUG nova.compute.provider_tree [None req-393edcfb-0595-4190-8ce0-4c14157dbbff tempest-ServerDiagnosticsNegativeTest-1508062492 tempest-ServerDiagnosticsNegativeTest-1508062492-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 706.981753] env[63088]: DEBUG oslo_concurrency.lockutils [req-240751f2-478d-4dba-a228-278dc60b1ffe req-b320b21b-37eb-4cd4-82b3-a99f537f24a7 service nova] Releasing lock "refresh_cache-02da88f0-7e3a-4eea-a88d-59a990c69f82" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 706.982175] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0b29042d-57c3-47a9-a926-f67d685a4740 tempest-TenantUsagesTestJSON-1054626739 tempest-TenantUsagesTestJSON-1054626739-project-member] Acquired lock "refresh_cache-02da88f0-7e3a-4eea-a88d-59a990c69f82" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 706.982364] env[63088]: DEBUG nova.network.neutron [None req-0b29042d-57c3-47a9-a926-f67d685a4740 tempest-TenantUsagesTestJSON-1054626739 tempest-TenantUsagesTestJSON-1054626739-project-member] [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 707.205034] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f237740d-6eaf-4423-b770-dd9775fe6fd5 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Releasing lock "refresh_cache-d9a4d198-4aee-4882-bee8-f6cde14745a2" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 707.205281] env[63088]: DEBUG nova.compute.manager [None req-f237740d-6eaf-4423-b770-dd9775fe6fd5 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 707.205467] env[63088]: DEBUG nova.compute.manager [None req-f237740d-6eaf-4423-b770-dd9775fe6fd5 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 707.205649] env[63088]: DEBUG nova.network.neutron [None req-f237740d-6eaf-4423-b770-dd9775fe6fd5 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 707.225874] env[63088]: DEBUG nova.network.neutron [None req-f237740d-6eaf-4423-b770-dd9775fe6fd5 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 707.456620] env[63088]: DEBUG nova.scheduler.client.report [None req-393edcfb-0595-4190-8ce0-4c14157dbbff tempest-ServerDiagnosticsNegativeTest-1508062492 tempest-ServerDiagnosticsNegativeTest-1508062492-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 707.499591] env[63088]: DEBUG nova.network.neutron [None req-0b29042d-57c3-47a9-a926-f67d685a4740 tempest-TenantUsagesTestJSON-1054626739 tempest-TenantUsagesTestJSON-1054626739-project-member] [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 707.542395] env[63088]: DEBUG nova.network.neutron [None req-0b29042d-57c3-47a9-a926-f67d685a4740 tempest-TenantUsagesTestJSON-1054626739 tempest-TenantUsagesTestJSON-1054626739-project-member] [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 707.731096] env[63088]: DEBUG nova.network.neutron [None req-f237740d-6eaf-4423-b770-dd9775fe6fd5 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 707.961486] env[63088]: DEBUG oslo_concurrency.lockutils [None req-393edcfb-0595-4190-8ce0-4c14157dbbff tempest-ServerDiagnosticsNegativeTest-1508062492 tempest-ServerDiagnosticsNegativeTest-1508062492-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.875s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 707.962096] env[63088]: ERROR nova.compute.manager [None req-393edcfb-0595-4190-8ce0-4c14157dbbff tempest-ServerDiagnosticsNegativeTest-1508062492 tempest-ServerDiagnosticsNegativeTest-1508062492-project-member] [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b4cabb9e-94ed-49f7-ab53-3e736fbbc48a, please check neutron logs for more information. [ 707.962096] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] Traceback (most recent call last): [ 707.962096] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 707.962096] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] self.driver.spawn(context, instance, image_meta, [ 707.962096] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 707.962096] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 707.962096] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 707.962096] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] vm_ref = self.build_virtual_machine(instance, [ 707.962096] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 707.962096] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] vif_infos = vmwarevif.get_vif_info(self._session, [ 707.962096] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 707.962472] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] for vif in network_info: [ 707.962472] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 707.962472] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] return self._sync_wrapper(fn, *args, **kwargs) [ 707.962472] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 707.962472] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] self.wait() [ 707.962472] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 707.962472] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] self[:] = self._gt.wait() [ 707.962472] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 707.962472] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] return self._exit_event.wait() [ 707.962472] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 707.962472] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] current.throw(*self._exc) [ 707.962472] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 707.962472] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] result = function(*args, **kwargs) [ 707.962855] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 707.962855] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] return func(*args, **kwargs) [ 707.962855] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 707.962855] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] raise e [ 707.962855] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 707.962855] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] nwinfo = self.network_api.allocate_for_instance( [ 707.962855] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 707.962855] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] created_port_ids = self._update_ports_for_instance( [ 707.962855] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 707.962855] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] with excutils.save_and_reraise_exception(): [ 707.962855] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 707.962855] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] self.force_reraise() [ 707.962855] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 707.964183] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] raise self.value [ 707.964183] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 707.964183] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] updated_port = self._update_port( [ 707.964183] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 707.964183] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] _ensure_no_port_binding_failure(port) [ 707.964183] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 707.964183] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] raise exception.PortBindingFailed(port_id=port['id']) [ 707.964183] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] nova.exception.PortBindingFailed: Binding failed for port b4cabb9e-94ed-49f7-ab53-3e736fbbc48a, please check neutron logs for more information. [ 707.964183] env[63088]: ERROR nova.compute.manager [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] [ 707.964183] env[63088]: DEBUG nova.compute.utils [None req-393edcfb-0595-4190-8ce0-4c14157dbbff tempest-ServerDiagnosticsNegativeTest-1508062492 tempest-ServerDiagnosticsNegativeTest-1508062492-project-member] [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] Binding failed for port b4cabb9e-94ed-49f7-ab53-3e736fbbc48a, please check neutron logs for more information. {{(pid=63088) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 707.964482] env[63088]: DEBUG oslo_concurrency.lockutils [None req-fcd5880e-5b8a-42fa-b638-13cd9836ac96 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.278s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 707.965812] env[63088]: INFO nova.compute.claims [None req-fcd5880e-5b8a-42fa-b638-13cd9836ac96 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 707.968401] env[63088]: DEBUG nova.compute.manager [None req-393edcfb-0595-4190-8ce0-4c14157dbbff tempest-ServerDiagnosticsNegativeTest-1508062492 tempest-ServerDiagnosticsNegativeTest-1508062492-project-member] [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] Build of instance 6eb3e767-ff7d-4760-a036-a55647afac0f was re-scheduled: Binding failed for port b4cabb9e-94ed-49f7-ab53-3e736fbbc48a, please check neutron logs for more information. {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 707.968859] env[63088]: DEBUG nova.compute.manager [None req-393edcfb-0595-4190-8ce0-4c14157dbbff tempest-ServerDiagnosticsNegativeTest-1508062492 tempest-ServerDiagnosticsNegativeTest-1508062492-project-member] [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] Unplugging VIFs for instance {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 707.969123] env[63088]: DEBUG oslo_concurrency.lockutils [None req-393edcfb-0595-4190-8ce0-4c14157dbbff tempest-ServerDiagnosticsNegativeTest-1508062492 tempest-ServerDiagnosticsNegativeTest-1508062492-project-member] Acquiring lock "refresh_cache-6eb3e767-ff7d-4760-a036-a55647afac0f" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 707.969273] env[63088]: DEBUG oslo_concurrency.lockutils [None req-393edcfb-0595-4190-8ce0-4c14157dbbff tempest-ServerDiagnosticsNegativeTest-1508062492 tempest-ServerDiagnosticsNegativeTest-1508062492-project-member] Acquired lock "refresh_cache-6eb3e767-ff7d-4760-a036-a55647afac0f" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 707.969433] env[63088]: DEBUG nova.network.neutron [None req-393edcfb-0595-4190-8ce0-4c14157dbbff tempest-ServerDiagnosticsNegativeTest-1508062492 tempest-ServerDiagnosticsNegativeTest-1508062492-project-member] [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 707.971940] env[63088]: DEBUG nova.compute.manager [req-2fb82271-4ba7-4853-9b41-15bb14c7014b req-d0bb13b8-fbc8-488b-917c-f5544221b4af service nova] [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] Received event network-vif-deleted-32cece70-208a-4d57-bcdf-d63d938be4d3 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 708.045242] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0b29042d-57c3-47a9-a926-f67d685a4740 tempest-TenantUsagesTestJSON-1054626739 tempest-TenantUsagesTestJSON-1054626739-project-member] Releasing lock "refresh_cache-02da88f0-7e3a-4eea-a88d-59a990c69f82" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 708.045675] env[63088]: DEBUG nova.compute.manager [None req-0b29042d-57c3-47a9-a926-f67d685a4740 tempest-TenantUsagesTestJSON-1054626739 tempest-TenantUsagesTestJSON-1054626739-project-member] [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 708.045864] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-0b29042d-57c3-47a9-a926-f67d685a4740 tempest-TenantUsagesTestJSON-1054626739 tempest-TenantUsagesTestJSON-1054626739-project-member] [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 708.046184] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-dcabdf38-1369-406a-8e4a-c7e3ec2509c5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.056258] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1f956a5-9c75-43b6-9daa-b3130c852276 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.077426] env[63088]: WARNING nova.virt.vmwareapi.vmops [None req-0b29042d-57c3-47a9-a926-f67d685a4740 tempest-TenantUsagesTestJSON-1054626739 tempest-TenantUsagesTestJSON-1054626739-project-member] [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 02da88f0-7e3a-4eea-a88d-59a990c69f82 could not be found. [ 708.078326] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-0b29042d-57c3-47a9-a926-f67d685a4740 tempest-TenantUsagesTestJSON-1054626739 tempest-TenantUsagesTestJSON-1054626739-project-member] [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 708.078326] env[63088]: INFO nova.compute.manager [None req-0b29042d-57c3-47a9-a926-f67d685a4740 tempest-TenantUsagesTestJSON-1054626739 tempest-TenantUsagesTestJSON-1054626739-project-member] [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] Took 0.03 seconds to destroy the instance on the hypervisor. [ 708.078326] env[63088]: DEBUG oslo.service.loopingcall [None req-0b29042d-57c3-47a9-a926-f67d685a4740 tempest-TenantUsagesTestJSON-1054626739 tempest-TenantUsagesTestJSON-1054626739-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 708.078326] env[63088]: DEBUG nova.compute.manager [-] [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 708.078533] env[63088]: DEBUG nova.network.neutron [-] [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 708.093285] env[63088]: DEBUG nova.network.neutron [-] [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 708.233442] env[63088]: INFO nova.compute.manager [None req-f237740d-6eaf-4423-b770-dd9775fe6fd5 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: d9a4d198-4aee-4882-bee8-f6cde14745a2] Took 1.03 seconds to deallocate network for instance. [ 708.488354] env[63088]: DEBUG nova.network.neutron [None req-393edcfb-0595-4190-8ce0-4c14157dbbff tempest-ServerDiagnosticsNegativeTest-1508062492 tempest-ServerDiagnosticsNegativeTest-1508062492-project-member] [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 708.541270] env[63088]: DEBUG nova.network.neutron [None req-393edcfb-0595-4190-8ce0-4c14157dbbff tempest-ServerDiagnosticsNegativeTest-1508062492 tempest-ServerDiagnosticsNegativeTest-1508062492-project-member] [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 708.595830] env[63088]: DEBUG nova.network.neutron [-] [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 709.043686] env[63088]: DEBUG oslo_concurrency.lockutils [None req-393edcfb-0595-4190-8ce0-4c14157dbbff tempest-ServerDiagnosticsNegativeTest-1508062492 tempest-ServerDiagnosticsNegativeTest-1508062492-project-member] Releasing lock "refresh_cache-6eb3e767-ff7d-4760-a036-a55647afac0f" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 709.043970] env[63088]: DEBUG nova.compute.manager [None req-393edcfb-0595-4190-8ce0-4c14157dbbff tempest-ServerDiagnosticsNegativeTest-1508062492 tempest-ServerDiagnosticsNegativeTest-1508062492-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 709.044100] env[63088]: DEBUG nova.compute.manager [None req-393edcfb-0595-4190-8ce0-4c14157dbbff tempest-ServerDiagnosticsNegativeTest-1508062492 tempest-ServerDiagnosticsNegativeTest-1508062492-project-member] [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 709.044271] env[63088]: DEBUG nova.network.neutron [None req-393edcfb-0595-4190-8ce0-4c14157dbbff tempest-ServerDiagnosticsNegativeTest-1508062492 tempest-ServerDiagnosticsNegativeTest-1508062492-project-member] [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 709.061025] env[63088]: DEBUG nova.network.neutron [None req-393edcfb-0595-4190-8ce0-4c14157dbbff tempest-ServerDiagnosticsNegativeTest-1508062492 tempest-ServerDiagnosticsNegativeTest-1508062492-project-member] [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 709.098407] env[63088]: INFO nova.compute.manager [-] [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] Took 1.02 seconds to deallocate network for instance. [ 709.100820] env[63088]: DEBUG nova.compute.claims [None req-0b29042d-57c3-47a9-a926-f67d685a4740 tempest-TenantUsagesTestJSON-1054626739 tempest-TenantUsagesTestJSON-1054626739-project-member] [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] Aborting claim: {{(pid=63088) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 709.100933] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0b29042d-57c3-47a9-a926-f67d685a4740 tempest-TenantUsagesTestJSON-1054626739 tempest-TenantUsagesTestJSON-1054626739-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 709.238699] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d2e5086-6f28-4b88-a0d3-b85b266853e2 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.246362] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-002ef7eb-9b91-43ca-b282-82dad6341036 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.279011] env[63088]: INFO nova.scheduler.client.report [None req-f237740d-6eaf-4423-b770-dd9775fe6fd5 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Deleted allocations for instance d9a4d198-4aee-4882-bee8-f6cde14745a2 [ 709.284938] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1cf6eb5-1599-4d7f-963e-aa4e73c4a360 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.295058] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be7c70a3-fcaf-4941-ab7d-3add7bbb5289 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.309061] env[63088]: DEBUG nova.compute.provider_tree [None req-fcd5880e-5b8a-42fa-b638-13cd9836ac96 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 709.563237] env[63088]: DEBUG nova.network.neutron [None req-393edcfb-0595-4190-8ce0-4c14157dbbff tempest-ServerDiagnosticsNegativeTest-1508062492 tempest-ServerDiagnosticsNegativeTest-1508062492-project-member] [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 709.790334] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f237740d-6eaf-4423-b770-dd9775fe6fd5 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Lock "d9a4d198-4aee-4882-bee8-f6cde14745a2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 115.785s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 709.812881] env[63088]: DEBUG nova.scheduler.client.report [None req-fcd5880e-5b8a-42fa-b638-13cd9836ac96 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 710.066605] env[63088]: INFO nova.compute.manager [None req-393edcfb-0595-4190-8ce0-4c14157dbbff tempest-ServerDiagnosticsNegativeTest-1508062492 tempest-ServerDiagnosticsNegativeTest-1508062492-project-member] [instance: 6eb3e767-ff7d-4760-a036-a55647afac0f] Took 1.02 seconds to deallocate network for instance. [ 710.293523] env[63088]: DEBUG nova.compute.manager [None req-a6077124-74e2-4777-9beb-9998c6275315 tempest-ServersTestBootFromVolume-2105176458 tempest-ServersTestBootFromVolume-2105176458-project-member] [instance: 407db84f-3322-42bc-baac-5528e856af8d] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 710.317049] env[63088]: DEBUG oslo_concurrency.lockutils [None req-fcd5880e-5b8a-42fa-b638-13cd9836ac96 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.353s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 710.317558] env[63088]: DEBUG nova.compute.manager [None req-fcd5880e-5b8a-42fa-b638-13cd9836ac96 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 710.320061] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6c63332a-ec25-45cc-9207-42e0be237d32 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.766s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 710.321798] env[63088]: INFO nova.compute.claims [None req-6c63332a-ec25-45cc-9207-42e0be237d32 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 710.815615] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a6077124-74e2-4777-9beb-9998c6275315 tempest-ServersTestBootFromVolume-2105176458 tempest-ServersTestBootFromVolume-2105176458-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 710.826056] env[63088]: DEBUG nova.compute.utils [None req-fcd5880e-5b8a-42fa-b638-13cd9836ac96 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 710.829493] env[63088]: DEBUG nova.compute.manager [None req-fcd5880e-5b8a-42fa-b638-13cd9836ac96 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 710.829660] env[63088]: DEBUG nova.network.neutron [None req-fcd5880e-5b8a-42fa-b638-13cd9836ac96 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 710.879036] env[63088]: DEBUG nova.policy [None req-fcd5880e-5b8a-42fa-b638-13cd9836ac96 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a01468bc41a84e8e8a52f49b6c9a0494', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8e0a144caecd490ebc144409af902624', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 711.096725] env[63088]: INFO nova.scheduler.client.report [None req-393edcfb-0595-4190-8ce0-4c14157dbbff tempest-ServerDiagnosticsNegativeTest-1508062492 tempest-ServerDiagnosticsNegativeTest-1508062492-project-member] Deleted allocations for instance 6eb3e767-ff7d-4760-a036-a55647afac0f [ 711.180895] env[63088]: DEBUG nova.network.neutron [None req-fcd5880e-5b8a-42fa-b638-13cd9836ac96 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] Successfully created port: 28873a04-8e37-4443-9cf4-585cce95276d {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 711.331159] env[63088]: DEBUG nova.compute.manager [None req-fcd5880e-5b8a-42fa-b638-13cd9836ac96 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 711.608642] env[63088]: DEBUG oslo_concurrency.lockutils [None req-393edcfb-0595-4190-8ce0-4c14157dbbff tempest-ServerDiagnosticsNegativeTest-1508062492 tempest-ServerDiagnosticsNegativeTest-1508062492-project-member] Lock "6eb3e767-ff7d-4760-a036-a55647afac0f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 117.010s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 711.625065] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6f0aa0b-0136-4494-bcc6-b0cb52fb4b83 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.633532] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8bef23c-4a2b-4817-b229-034a533b927b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.666671] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b8ccbe1-f1f3-44ab-b3d5-c58c616ffd74 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.674268] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1f22264-e5f0-44e5-ad1d-3d4b3b8c377a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.690032] env[63088]: DEBUG nova.compute.provider_tree [None req-6c63332a-ec25-45cc-9207-42e0be237d32 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 712.081723] env[63088]: DEBUG nova.compute.manager [req-971a3e37-904e-4aea-b335-496b4ab4de76 req-8cbd173b-1777-47d8-8ca6-fdb6194ee5f8 service nova] [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] Received event network-changed-28873a04-8e37-4443-9cf4-585cce95276d {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 712.081960] env[63088]: DEBUG nova.compute.manager [req-971a3e37-904e-4aea-b335-496b4ab4de76 req-8cbd173b-1777-47d8-8ca6-fdb6194ee5f8 service nova] [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] Refreshing instance network info cache due to event network-changed-28873a04-8e37-4443-9cf4-585cce95276d. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 712.082136] env[63088]: DEBUG oslo_concurrency.lockutils [req-971a3e37-904e-4aea-b335-496b4ab4de76 req-8cbd173b-1777-47d8-8ca6-fdb6194ee5f8 service nova] Acquiring lock "refresh_cache-c67b276f-07db-4cde-9f1c-efccc4d11c7d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 712.082286] env[63088]: DEBUG oslo_concurrency.lockutils [req-971a3e37-904e-4aea-b335-496b4ab4de76 req-8cbd173b-1777-47d8-8ca6-fdb6194ee5f8 service nova] Acquired lock "refresh_cache-c67b276f-07db-4cde-9f1c-efccc4d11c7d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 712.082583] env[63088]: DEBUG nova.network.neutron [req-971a3e37-904e-4aea-b335-496b4ab4de76 req-8cbd173b-1777-47d8-8ca6-fdb6194ee5f8 service nova] [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] Refreshing network info cache for port 28873a04-8e37-4443-9cf4-585cce95276d {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 712.113109] env[63088]: DEBUG nova.compute.manager [None req-3b417ae5-c397-4c95-9150-090b3a6260b8 tempest-FloatingIPsAssociationTestJSON-911050413 tempest-FloatingIPsAssociationTestJSON-911050413-project-member] [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 712.195847] env[63088]: DEBUG nova.scheduler.client.report [None req-6c63332a-ec25-45cc-9207-42e0be237d32 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 712.250984] env[63088]: ERROR nova.compute.manager [None req-fcd5880e-5b8a-42fa-b638-13cd9836ac96 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 28873a04-8e37-4443-9cf4-585cce95276d, please check neutron logs for more information. [ 712.250984] env[63088]: ERROR nova.compute.manager Traceback (most recent call last): [ 712.250984] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 712.250984] env[63088]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 712.250984] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 712.250984] env[63088]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 712.250984] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 712.250984] env[63088]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 712.250984] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 712.250984] env[63088]: ERROR nova.compute.manager self.force_reraise() [ 712.250984] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 712.250984] env[63088]: ERROR nova.compute.manager raise self.value [ 712.250984] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 712.250984] env[63088]: ERROR nova.compute.manager updated_port = self._update_port( [ 712.250984] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 712.250984] env[63088]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 712.251600] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 712.251600] env[63088]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 712.251600] env[63088]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 28873a04-8e37-4443-9cf4-585cce95276d, please check neutron logs for more information. [ 712.251600] env[63088]: ERROR nova.compute.manager [ 712.251600] env[63088]: Traceback (most recent call last): [ 712.251600] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 712.251600] env[63088]: listener.cb(fileno) [ 712.251600] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 712.251600] env[63088]: result = function(*args, **kwargs) [ 712.251600] env[63088]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 712.251600] env[63088]: return func(*args, **kwargs) [ 712.251600] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 712.251600] env[63088]: raise e [ 712.251600] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 712.251600] env[63088]: nwinfo = self.network_api.allocate_for_instance( [ 712.251600] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 712.251600] env[63088]: created_port_ids = self._update_ports_for_instance( [ 712.251600] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 712.251600] env[63088]: with excutils.save_and_reraise_exception(): [ 712.251600] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 712.251600] env[63088]: self.force_reraise() [ 712.251600] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 712.251600] env[63088]: raise self.value [ 712.251600] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 712.251600] env[63088]: updated_port = self._update_port( [ 712.251600] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 712.251600] env[63088]: _ensure_no_port_binding_failure(port) [ 712.251600] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 712.251600] env[63088]: raise exception.PortBindingFailed(port_id=port['id']) [ 712.252761] env[63088]: nova.exception.PortBindingFailed: Binding failed for port 28873a04-8e37-4443-9cf4-585cce95276d, please check neutron logs for more information. [ 712.252761] env[63088]: Removing descriptor: 17 [ 712.345106] env[63088]: DEBUG nova.compute.manager [None req-fcd5880e-5b8a-42fa-b638-13cd9836ac96 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 712.371182] env[63088]: DEBUG nova.virt.hardware [None req-fcd5880e-5b8a-42fa-b638-13cd9836ac96 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:33Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 712.371440] env[63088]: DEBUG nova.virt.hardware [None req-fcd5880e-5b8a-42fa-b638-13cd9836ac96 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 712.371595] env[63088]: DEBUG nova.virt.hardware [None req-fcd5880e-5b8a-42fa-b638-13cd9836ac96 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 712.371771] env[63088]: DEBUG nova.virt.hardware [None req-fcd5880e-5b8a-42fa-b638-13cd9836ac96 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 712.371913] env[63088]: DEBUG nova.virt.hardware [None req-fcd5880e-5b8a-42fa-b638-13cd9836ac96 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 712.372069] env[63088]: DEBUG nova.virt.hardware [None req-fcd5880e-5b8a-42fa-b638-13cd9836ac96 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 712.372282] env[63088]: DEBUG nova.virt.hardware [None req-fcd5880e-5b8a-42fa-b638-13cd9836ac96 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 712.372439] env[63088]: DEBUG nova.virt.hardware [None req-fcd5880e-5b8a-42fa-b638-13cd9836ac96 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 712.372604] env[63088]: DEBUG nova.virt.hardware [None req-fcd5880e-5b8a-42fa-b638-13cd9836ac96 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 712.372763] env[63088]: DEBUG nova.virt.hardware [None req-fcd5880e-5b8a-42fa-b638-13cd9836ac96 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 712.372933] env[63088]: DEBUG nova.virt.hardware [None req-fcd5880e-5b8a-42fa-b638-13cd9836ac96 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 712.373803] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-caa1a884-7fa3-4cbb-b063-337aac09240b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.382144] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10caf4df-be89-48e5-a8f2-c649faa958ee {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.396573] env[63088]: ERROR nova.compute.manager [None req-fcd5880e-5b8a-42fa-b638-13cd9836ac96 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 28873a04-8e37-4443-9cf4-585cce95276d, please check neutron logs for more information. [ 712.396573] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] Traceback (most recent call last): [ 712.396573] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 712.396573] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] yield resources [ 712.396573] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 712.396573] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] self.driver.spawn(context, instance, image_meta, [ 712.396573] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 712.396573] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 712.396573] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 712.396573] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] vm_ref = self.build_virtual_machine(instance, [ 712.396573] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 712.397010] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] vif_infos = vmwarevif.get_vif_info(self._session, [ 712.397010] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 712.397010] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] for vif in network_info: [ 712.397010] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 712.397010] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] return self._sync_wrapper(fn, *args, **kwargs) [ 712.397010] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 712.397010] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] self.wait() [ 712.397010] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 712.397010] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] self[:] = self._gt.wait() [ 712.397010] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 712.397010] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] return self._exit_event.wait() [ 712.397010] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 712.397010] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] current.throw(*self._exc) [ 712.397427] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 712.397427] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] result = function(*args, **kwargs) [ 712.397427] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 712.397427] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] return func(*args, **kwargs) [ 712.397427] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 712.397427] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] raise e [ 712.397427] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 712.397427] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] nwinfo = self.network_api.allocate_for_instance( [ 712.397427] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 712.397427] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] created_port_ids = self._update_ports_for_instance( [ 712.397427] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 712.397427] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] with excutils.save_and_reraise_exception(): [ 712.397427] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 712.397843] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] self.force_reraise() [ 712.397843] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 712.397843] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] raise self.value [ 712.397843] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 712.397843] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] updated_port = self._update_port( [ 712.397843] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 712.397843] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] _ensure_no_port_binding_failure(port) [ 712.397843] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 712.397843] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] raise exception.PortBindingFailed(port_id=port['id']) [ 712.397843] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] nova.exception.PortBindingFailed: Binding failed for port 28873a04-8e37-4443-9cf4-585cce95276d, please check neutron logs for more information. [ 712.397843] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] [ 712.397843] env[63088]: INFO nova.compute.manager [None req-fcd5880e-5b8a-42fa-b638-13cd9836ac96 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] Terminating instance [ 712.399115] env[63088]: DEBUG oslo_concurrency.lockutils [None req-fcd5880e-5b8a-42fa-b638-13cd9836ac96 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Acquiring lock "refresh_cache-c67b276f-07db-4cde-9f1c-efccc4d11c7d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 712.607755] env[63088]: DEBUG nova.network.neutron [req-971a3e37-904e-4aea-b335-496b4ab4de76 req-8cbd173b-1777-47d8-8ca6-fdb6194ee5f8 service nova] [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 712.631355] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3b417ae5-c397-4c95-9150-090b3a6260b8 tempest-FloatingIPsAssociationTestJSON-911050413 tempest-FloatingIPsAssociationTestJSON-911050413-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 712.697682] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6c63332a-ec25-45cc-9207-42e0be237d32 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.378s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 712.698201] env[63088]: DEBUG nova.compute.manager [None req-6c63332a-ec25-45cc-9207-42e0be237d32 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 712.706549] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0c6bc062-cb90-4d50-aabf-8c0503771f7e tempest-InstanceActionsNegativeTestJSON-1719444358 tempest-InstanceActionsNegativeTestJSON-1719444358-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.254s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 712.708008] env[63088]: INFO nova.compute.claims [None req-0c6bc062-cb90-4d50-aabf-8c0503771f7e tempest-InstanceActionsNegativeTestJSON-1719444358 tempest-InstanceActionsNegativeTestJSON-1719444358-project-member] [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 712.714034] env[63088]: DEBUG nova.network.neutron [req-971a3e37-904e-4aea-b335-496b4ab4de76 req-8cbd173b-1777-47d8-8ca6-fdb6194ee5f8 service nova] [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 713.214026] env[63088]: DEBUG nova.compute.utils [None req-6c63332a-ec25-45cc-9207-42e0be237d32 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 713.218211] env[63088]: DEBUG nova.compute.manager [None req-6c63332a-ec25-45cc-9207-42e0be237d32 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 713.218405] env[63088]: DEBUG nova.network.neutron [None req-6c63332a-ec25-45cc-9207-42e0be237d32 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 713.221367] env[63088]: DEBUG oslo_concurrency.lockutils [req-971a3e37-904e-4aea-b335-496b4ab4de76 req-8cbd173b-1777-47d8-8ca6-fdb6194ee5f8 service nova] Releasing lock "refresh_cache-c67b276f-07db-4cde-9f1c-efccc4d11c7d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 713.221888] env[63088]: DEBUG oslo_concurrency.lockutils [None req-fcd5880e-5b8a-42fa-b638-13cd9836ac96 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Acquired lock "refresh_cache-c67b276f-07db-4cde-9f1c-efccc4d11c7d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 713.221964] env[63088]: DEBUG nova.network.neutron [None req-fcd5880e-5b8a-42fa-b638-13cd9836ac96 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 713.270085] env[63088]: DEBUG nova.policy [None req-6c63332a-ec25-45cc-9207-42e0be237d32 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '93f1d334e71a4a99ba300c3ee6d70243', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'aa3d24a1a6c0430985fd80365d986ee1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 713.562528] env[63088]: DEBUG nova.network.neutron [None req-6c63332a-ec25-45cc-9207-42e0be237d32 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] Successfully created port: c47ac4c5-a4a8-4766-bf26-05d24b003141 {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 713.719551] env[63088]: DEBUG nova.compute.manager [None req-6c63332a-ec25-45cc-9207-42e0be237d32 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 713.750460] env[63088]: DEBUG nova.network.neutron [None req-fcd5880e-5b8a-42fa-b638-13cd9836ac96 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 713.889539] env[63088]: DEBUG nova.network.neutron [None req-fcd5880e-5b8a-42fa-b638-13cd9836ac96 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 714.044230] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b889417-b91d-4759-be5f-8301a02f6506 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.052196] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86418bc5-4471-431a-bb01-cc7bbba54a0a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.083250] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17e1d16c-3806-45f2-8f19-6079cd9f5ea3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.090877] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d01914f5-aeee-474e-b890-d1ff936165cd {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.104326] env[63088]: DEBUG nova.compute.provider_tree [None req-0c6bc062-cb90-4d50-aabf-8c0503771f7e tempest-InstanceActionsNegativeTestJSON-1719444358 tempest-InstanceActionsNegativeTestJSON-1719444358-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 714.147543] env[63088]: DEBUG nova.compute.manager [req-f947bec7-9df4-4777-90f1-bf1c47aa2f50 req-7cd61fd7-46b8-48c9-83b1-22d444f4c072 service nova] [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] Received event network-vif-deleted-28873a04-8e37-4443-9cf4-585cce95276d {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 714.393451] env[63088]: DEBUG oslo_concurrency.lockutils [None req-fcd5880e-5b8a-42fa-b638-13cd9836ac96 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Releasing lock "refresh_cache-c67b276f-07db-4cde-9f1c-efccc4d11c7d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 714.393451] env[63088]: DEBUG nova.compute.manager [None req-fcd5880e-5b8a-42fa-b638-13cd9836ac96 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 714.393844] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-fcd5880e-5b8a-42fa-b638-13cd9836ac96 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 714.393844] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-40c5f038-7e6a-4180-a807-73edc3c5a40c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.406314] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81bfbaf1-1fc5-4dca-a9a5-1af19e672837 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.431364] env[63088]: WARNING nova.virt.vmwareapi.vmops [None req-fcd5880e-5b8a-42fa-b638-13cd9836ac96 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c67b276f-07db-4cde-9f1c-efccc4d11c7d could not be found. [ 714.431364] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-fcd5880e-5b8a-42fa-b638-13cd9836ac96 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 714.431364] env[63088]: INFO nova.compute.manager [None req-fcd5880e-5b8a-42fa-b638-13cd9836ac96 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] Took 0.04 seconds to destroy the instance on the hypervisor. [ 714.431364] env[63088]: DEBUG oslo.service.loopingcall [None req-fcd5880e-5b8a-42fa-b638-13cd9836ac96 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 714.431364] env[63088]: DEBUG nova.compute.manager [-] [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 714.431364] env[63088]: DEBUG nova.network.neutron [-] [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 714.464708] env[63088]: DEBUG nova.network.neutron [-] [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 714.570853] env[63088]: ERROR nova.compute.manager [None req-6c63332a-ec25-45cc-9207-42e0be237d32 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c47ac4c5-a4a8-4766-bf26-05d24b003141, please check neutron logs for more information. [ 714.570853] env[63088]: ERROR nova.compute.manager Traceback (most recent call last): [ 714.570853] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 714.570853] env[63088]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 714.570853] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 714.570853] env[63088]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 714.570853] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 714.570853] env[63088]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 714.570853] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 714.570853] env[63088]: ERROR nova.compute.manager self.force_reraise() [ 714.570853] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 714.570853] env[63088]: ERROR nova.compute.manager raise self.value [ 714.570853] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 714.570853] env[63088]: ERROR nova.compute.manager updated_port = self._update_port( [ 714.570853] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 714.570853] env[63088]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 714.571494] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 714.571494] env[63088]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 714.571494] env[63088]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c47ac4c5-a4a8-4766-bf26-05d24b003141, please check neutron logs for more information. [ 714.571494] env[63088]: ERROR nova.compute.manager [ 714.571494] env[63088]: Traceback (most recent call last): [ 714.571494] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 714.571494] env[63088]: listener.cb(fileno) [ 714.571494] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 714.571494] env[63088]: result = function(*args, **kwargs) [ 714.571494] env[63088]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 714.571494] env[63088]: return func(*args, **kwargs) [ 714.571494] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 714.571494] env[63088]: raise e [ 714.571494] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 714.571494] env[63088]: nwinfo = self.network_api.allocate_for_instance( [ 714.571494] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 714.571494] env[63088]: created_port_ids = self._update_ports_for_instance( [ 714.571494] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 714.571494] env[63088]: with excutils.save_and_reraise_exception(): [ 714.571494] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 714.571494] env[63088]: self.force_reraise() [ 714.571494] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 714.571494] env[63088]: raise self.value [ 714.571494] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 714.571494] env[63088]: updated_port = self._update_port( [ 714.571494] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 714.571494] env[63088]: _ensure_no_port_binding_failure(port) [ 714.571494] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 714.571494] env[63088]: raise exception.PortBindingFailed(port_id=port['id']) [ 714.572466] env[63088]: nova.exception.PortBindingFailed: Binding failed for port c47ac4c5-a4a8-4766-bf26-05d24b003141, please check neutron logs for more information. [ 714.572466] env[63088]: Removing descriptor: 17 [ 714.607084] env[63088]: DEBUG nova.scheduler.client.report [None req-0c6bc062-cb90-4d50-aabf-8c0503771f7e tempest-InstanceActionsNegativeTestJSON-1719444358 tempest-InstanceActionsNegativeTestJSON-1719444358-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 714.731907] env[63088]: DEBUG nova.compute.manager [None req-6c63332a-ec25-45cc-9207-42e0be237d32 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 714.756286] env[63088]: DEBUG nova.virt.hardware [None req-6c63332a-ec25-45cc-9207-42e0be237d32 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 714.756598] env[63088]: DEBUG nova.virt.hardware [None req-6c63332a-ec25-45cc-9207-42e0be237d32 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 714.756754] env[63088]: DEBUG nova.virt.hardware [None req-6c63332a-ec25-45cc-9207-42e0be237d32 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 714.756931] env[63088]: DEBUG nova.virt.hardware [None req-6c63332a-ec25-45cc-9207-42e0be237d32 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 714.757117] env[63088]: DEBUG nova.virt.hardware [None req-6c63332a-ec25-45cc-9207-42e0be237d32 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 714.757272] env[63088]: DEBUG nova.virt.hardware [None req-6c63332a-ec25-45cc-9207-42e0be237d32 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 714.757473] env[63088]: DEBUG nova.virt.hardware [None req-6c63332a-ec25-45cc-9207-42e0be237d32 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 714.757629] env[63088]: DEBUG nova.virt.hardware [None req-6c63332a-ec25-45cc-9207-42e0be237d32 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 714.757790] env[63088]: DEBUG nova.virt.hardware [None req-6c63332a-ec25-45cc-9207-42e0be237d32 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 714.757950] env[63088]: DEBUG nova.virt.hardware [None req-6c63332a-ec25-45cc-9207-42e0be237d32 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 714.758151] env[63088]: DEBUG nova.virt.hardware [None req-6c63332a-ec25-45cc-9207-42e0be237d32 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 714.759014] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65176b08-b234-405b-b01f-6d9652bd9f93 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.768882] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7f4ac7b-45f8-48c1-9615-193c5ff14f08 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.784476] env[63088]: ERROR nova.compute.manager [None req-6c63332a-ec25-45cc-9207-42e0be237d32 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c47ac4c5-a4a8-4766-bf26-05d24b003141, please check neutron logs for more information. [ 714.784476] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] Traceback (most recent call last): [ 714.784476] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 714.784476] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] yield resources [ 714.784476] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 714.784476] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] self.driver.spawn(context, instance, image_meta, [ 714.784476] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 714.784476] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 714.784476] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 714.784476] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] vm_ref = self.build_virtual_machine(instance, [ 714.784476] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 714.784891] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] vif_infos = vmwarevif.get_vif_info(self._session, [ 714.784891] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 714.784891] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] for vif in network_info: [ 714.784891] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 714.784891] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] return self._sync_wrapper(fn, *args, **kwargs) [ 714.784891] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 714.784891] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] self.wait() [ 714.784891] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 714.784891] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] self[:] = self._gt.wait() [ 714.784891] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 714.784891] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] return self._exit_event.wait() [ 714.784891] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 714.784891] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] current.throw(*self._exc) [ 714.785273] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 714.785273] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] result = function(*args, **kwargs) [ 714.785273] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 714.785273] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] return func(*args, **kwargs) [ 714.785273] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 714.785273] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] raise e [ 714.785273] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 714.785273] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] nwinfo = self.network_api.allocate_for_instance( [ 714.785273] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 714.785273] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] created_port_ids = self._update_ports_for_instance( [ 714.785273] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 714.785273] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] with excutils.save_and_reraise_exception(): [ 714.785273] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 714.785667] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] self.force_reraise() [ 714.785667] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 714.785667] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] raise self.value [ 714.785667] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 714.785667] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] updated_port = self._update_port( [ 714.785667] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 714.785667] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] _ensure_no_port_binding_failure(port) [ 714.785667] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 714.785667] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] raise exception.PortBindingFailed(port_id=port['id']) [ 714.785667] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] nova.exception.PortBindingFailed: Binding failed for port c47ac4c5-a4a8-4766-bf26-05d24b003141, please check neutron logs for more information. [ 714.785667] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] [ 714.785667] env[63088]: INFO nova.compute.manager [None req-6c63332a-ec25-45cc-9207-42e0be237d32 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] Terminating instance [ 714.788078] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6c63332a-ec25-45cc-9207-42e0be237d32 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquiring lock "refresh_cache-1219fe86-f218-4a2b-84ba-4a377829386d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 714.788078] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6c63332a-ec25-45cc-9207-42e0be237d32 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquired lock "refresh_cache-1219fe86-f218-4a2b-84ba-4a377829386d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 714.788078] env[63088]: DEBUG nova.network.neutron [None req-6c63332a-ec25-45cc-9207-42e0be237d32 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 714.965724] env[63088]: DEBUG nova.network.neutron [-] [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 715.117605] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0c6bc062-cb90-4d50-aabf-8c0503771f7e tempest-InstanceActionsNegativeTestJSON-1719444358 tempest-InstanceActionsNegativeTestJSON-1719444358-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.411s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 715.118132] env[63088]: DEBUG nova.compute.manager [None req-0c6bc062-cb90-4d50-aabf-8c0503771f7e tempest-InstanceActionsNegativeTestJSON-1719444358 tempest-InstanceActionsNegativeTestJSON-1719444358-project-member] [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 715.121583] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a2a0ac90-0e7a-45d1-a070-f7ee6be01d68 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.676s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 715.122293] env[63088]: INFO nova.compute.claims [None req-a2a0ac90-0e7a-45d1-a070-f7ee6be01d68 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 715.324224] env[63088]: DEBUG nova.network.neutron [None req-6c63332a-ec25-45cc-9207-42e0be237d32 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 715.462111] env[63088]: DEBUG nova.network.neutron [None req-6c63332a-ec25-45cc-9207-42e0be237d32 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 715.470169] env[63088]: INFO nova.compute.manager [-] [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] Took 1.04 seconds to deallocate network for instance. [ 715.472488] env[63088]: DEBUG nova.compute.claims [None req-fcd5880e-5b8a-42fa-b638-13cd9836ac96 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] Aborting claim: {{(pid=63088) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 715.472826] env[63088]: DEBUG oslo_concurrency.lockutils [None req-fcd5880e-5b8a-42fa-b638-13cd9836ac96 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 715.626813] env[63088]: DEBUG nova.compute.utils [None req-0c6bc062-cb90-4d50-aabf-8c0503771f7e tempest-InstanceActionsNegativeTestJSON-1719444358 tempest-InstanceActionsNegativeTestJSON-1719444358-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 715.630449] env[63088]: DEBUG nova.compute.manager [None req-0c6bc062-cb90-4d50-aabf-8c0503771f7e tempest-InstanceActionsNegativeTestJSON-1719444358 tempest-InstanceActionsNegativeTestJSON-1719444358-project-member] [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 715.630622] env[63088]: DEBUG nova.network.neutron [None req-0c6bc062-cb90-4d50-aabf-8c0503771f7e tempest-InstanceActionsNegativeTestJSON-1719444358 tempest-InstanceActionsNegativeTestJSON-1719444358-project-member] [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 715.693600] env[63088]: DEBUG nova.policy [None req-0c6bc062-cb90-4d50-aabf-8c0503771f7e tempest-InstanceActionsNegativeTestJSON-1719444358 tempest-InstanceActionsNegativeTestJSON-1719444358-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '82684237e9964fbfa8ca2097f8c9314a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '386a1c929f36470b838284fcbb2bd7d2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 715.967354] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6c63332a-ec25-45cc-9207-42e0be237d32 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Releasing lock "refresh_cache-1219fe86-f218-4a2b-84ba-4a377829386d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 715.967783] env[63088]: DEBUG nova.compute.manager [None req-6c63332a-ec25-45cc-9207-42e0be237d32 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 715.968017] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-6c63332a-ec25-45cc-9207-42e0be237d32 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 715.968339] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0c64603c-2d2e-400c-b20c-3a723ef8cdab {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.979157] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c52be4f-6493-4de7-a4cc-f290ad042859 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.992354] env[63088]: DEBUG nova.network.neutron [None req-0c6bc062-cb90-4d50-aabf-8c0503771f7e tempest-InstanceActionsNegativeTestJSON-1719444358 tempest-InstanceActionsNegativeTestJSON-1719444358-project-member] [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] Successfully created port: 9c453a07-3eda-4075-81ce-a4516dc00de7 {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 716.002297] env[63088]: WARNING nova.virt.vmwareapi.vmops [None req-6c63332a-ec25-45cc-9207-42e0be237d32 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1219fe86-f218-4a2b-84ba-4a377829386d could not be found. [ 716.002517] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-6c63332a-ec25-45cc-9207-42e0be237d32 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 716.002702] env[63088]: INFO nova.compute.manager [None req-6c63332a-ec25-45cc-9207-42e0be237d32 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] Took 0.03 seconds to destroy the instance on the hypervisor. [ 716.002954] env[63088]: DEBUG oslo.service.loopingcall [None req-6c63332a-ec25-45cc-9207-42e0be237d32 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 716.003883] env[63088]: DEBUG nova.compute.manager [-] [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 716.003883] env[63088]: DEBUG nova.network.neutron [-] [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 716.026164] env[63088]: DEBUG nova.network.neutron [-] [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 716.131443] env[63088]: DEBUG nova.compute.manager [None req-0c6bc062-cb90-4d50-aabf-8c0503771f7e tempest-InstanceActionsNegativeTestJSON-1719444358 tempest-InstanceActionsNegativeTestJSON-1719444358-project-member] [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 716.191053] env[63088]: DEBUG nova.compute.manager [req-5298851f-a6eb-480d-986a-2b5df18bdab2 req-9f23c420-e345-407d-a028-ee9861cffdad service nova] [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] Received event network-changed-c47ac4c5-a4a8-4766-bf26-05d24b003141 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 716.191053] env[63088]: DEBUG nova.compute.manager [req-5298851f-a6eb-480d-986a-2b5df18bdab2 req-9f23c420-e345-407d-a028-ee9861cffdad service nova] [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] Refreshing instance network info cache due to event network-changed-c47ac4c5-a4a8-4766-bf26-05d24b003141. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 716.191053] env[63088]: DEBUG oslo_concurrency.lockutils [req-5298851f-a6eb-480d-986a-2b5df18bdab2 req-9f23c420-e345-407d-a028-ee9861cffdad service nova] Acquiring lock "refresh_cache-1219fe86-f218-4a2b-84ba-4a377829386d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 716.191053] env[63088]: DEBUG oslo_concurrency.lockutils [req-5298851f-a6eb-480d-986a-2b5df18bdab2 req-9f23c420-e345-407d-a028-ee9861cffdad service nova] Acquired lock "refresh_cache-1219fe86-f218-4a2b-84ba-4a377829386d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 716.191053] env[63088]: DEBUG nova.network.neutron [req-5298851f-a6eb-480d-986a-2b5df18bdab2 req-9f23c420-e345-407d-a028-ee9861cffdad service nova] [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] Refreshing network info cache for port c47ac4c5-a4a8-4766-bf26-05d24b003141 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 716.476963] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26f0d6e0-0111-481b-bfcf-2a0e4087cc4d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.485480] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43343e9c-df86-4258-8c58-f476b4196e22 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.532248] env[63088]: DEBUG nova.network.neutron [-] [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 716.536168] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c95dc74-55e9-496c-9d28-be9028f9337f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.545708] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b67032ef-36a2-4250-b9bb-37d4e50daf3e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.560595] env[63088]: DEBUG nova.compute.provider_tree [None req-a2a0ac90-0e7a-45d1-a070-f7ee6be01d68 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 716.715303] env[63088]: DEBUG nova.network.neutron [req-5298851f-a6eb-480d-986a-2b5df18bdab2 req-9f23c420-e345-407d-a028-ee9861cffdad service nova] [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 716.800015] env[63088]: DEBUG nova.network.neutron [req-5298851f-a6eb-480d-986a-2b5df18bdab2 req-9f23c420-e345-407d-a028-ee9861cffdad service nova] [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 716.990234] env[63088]: DEBUG oslo_concurrency.lockutils [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquiring lock "11ec987f-9908-4366-81ae-cbfe4de67b2d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 716.990449] env[63088]: DEBUG oslo_concurrency.lockutils [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "11ec987f-9908-4366-81ae-cbfe4de67b2d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 717.035935] env[63088]: INFO nova.compute.manager [-] [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] Took 1.03 seconds to deallocate network for instance. [ 717.039202] env[63088]: DEBUG nova.compute.claims [None req-6c63332a-ec25-45cc-9207-42e0be237d32 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] Aborting claim: {{(pid=63088) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 717.039202] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6c63332a-ec25-45cc-9207-42e0be237d32 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 717.064015] env[63088]: DEBUG nova.scheduler.client.report [None req-a2a0ac90-0e7a-45d1-a070-f7ee6be01d68 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 717.147317] env[63088]: DEBUG nova.compute.manager [None req-0c6bc062-cb90-4d50-aabf-8c0503771f7e tempest-InstanceActionsNegativeTestJSON-1719444358 tempest-InstanceActionsNegativeTestJSON-1719444358-project-member] [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 717.168547] env[63088]: ERROR nova.compute.manager [None req-0c6bc062-cb90-4d50-aabf-8c0503771f7e tempest-InstanceActionsNegativeTestJSON-1719444358 tempest-InstanceActionsNegativeTestJSON-1719444358-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9c453a07-3eda-4075-81ce-a4516dc00de7, please check neutron logs for more information. [ 717.168547] env[63088]: ERROR nova.compute.manager Traceback (most recent call last): [ 717.168547] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 717.168547] env[63088]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 717.168547] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 717.168547] env[63088]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 717.168547] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 717.168547] env[63088]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 717.168547] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 717.168547] env[63088]: ERROR nova.compute.manager self.force_reraise() [ 717.168547] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 717.168547] env[63088]: ERROR nova.compute.manager raise self.value [ 717.168547] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 717.168547] env[63088]: ERROR nova.compute.manager updated_port = self._update_port( [ 717.168547] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 717.168547] env[63088]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 717.169105] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 717.169105] env[63088]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 717.169105] env[63088]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9c453a07-3eda-4075-81ce-a4516dc00de7, please check neutron logs for more information. [ 717.169105] env[63088]: ERROR nova.compute.manager [ 717.169105] env[63088]: Traceback (most recent call last): [ 717.169105] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 717.169105] env[63088]: listener.cb(fileno) [ 717.169105] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 717.169105] env[63088]: result = function(*args, **kwargs) [ 717.169105] env[63088]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 717.169105] env[63088]: return func(*args, **kwargs) [ 717.169105] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 717.169105] env[63088]: raise e [ 717.169105] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 717.169105] env[63088]: nwinfo = self.network_api.allocate_for_instance( [ 717.169105] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 717.169105] env[63088]: created_port_ids = self._update_ports_for_instance( [ 717.169105] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 717.169105] env[63088]: with excutils.save_and_reraise_exception(): [ 717.169105] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 717.169105] env[63088]: self.force_reraise() [ 717.169105] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 717.169105] env[63088]: raise self.value [ 717.169105] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 717.169105] env[63088]: updated_port = self._update_port( [ 717.169105] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 717.169105] env[63088]: _ensure_no_port_binding_failure(port) [ 717.169105] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 717.169105] env[63088]: raise exception.PortBindingFailed(port_id=port['id']) [ 717.169850] env[63088]: nova.exception.PortBindingFailed: Binding failed for port 9c453a07-3eda-4075-81ce-a4516dc00de7, please check neutron logs for more information. [ 717.169850] env[63088]: Removing descriptor: 17 [ 717.173343] env[63088]: DEBUG nova.virt.hardware [None req-0c6bc062-cb90-4d50-aabf-8c0503771f7e tempest-InstanceActionsNegativeTestJSON-1719444358 tempest-InstanceActionsNegativeTestJSON-1719444358-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 717.173584] env[63088]: DEBUG nova.virt.hardware [None req-0c6bc062-cb90-4d50-aabf-8c0503771f7e tempest-InstanceActionsNegativeTestJSON-1719444358 tempest-InstanceActionsNegativeTestJSON-1719444358-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 717.173739] env[63088]: DEBUG nova.virt.hardware [None req-0c6bc062-cb90-4d50-aabf-8c0503771f7e tempest-InstanceActionsNegativeTestJSON-1719444358 tempest-InstanceActionsNegativeTestJSON-1719444358-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 717.173917] env[63088]: DEBUG nova.virt.hardware [None req-0c6bc062-cb90-4d50-aabf-8c0503771f7e tempest-InstanceActionsNegativeTestJSON-1719444358 tempest-InstanceActionsNegativeTestJSON-1719444358-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 717.174076] env[63088]: DEBUG nova.virt.hardware [None req-0c6bc062-cb90-4d50-aabf-8c0503771f7e tempest-InstanceActionsNegativeTestJSON-1719444358 tempest-InstanceActionsNegativeTestJSON-1719444358-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 717.174230] env[63088]: DEBUG nova.virt.hardware [None req-0c6bc062-cb90-4d50-aabf-8c0503771f7e tempest-InstanceActionsNegativeTestJSON-1719444358 tempest-InstanceActionsNegativeTestJSON-1719444358-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 717.174438] env[63088]: DEBUG nova.virt.hardware [None req-0c6bc062-cb90-4d50-aabf-8c0503771f7e tempest-InstanceActionsNegativeTestJSON-1719444358 tempest-InstanceActionsNegativeTestJSON-1719444358-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 717.174598] env[63088]: DEBUG nova.virt.hardware [None req-0c6bc062-cb90-4d50-aabf-8c0503771f7e tempest-InstanceActionsNegativeTestJSON-1719444358 tempest-InstanceActionsNegativeTestJSON-1719444358-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 717.174762] env[63088]: DEBUG nova.virt.hardware [None req-0c6bc062-cb90-4d50-aabf-8c0503771f7e tempest-InstanceActionsNegativeTestJSON-1719444358 tempest-InstanceActionsNegativeTestJSON-1719444358-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 717.174923] env[63088]: DEBUG nova.virt.hardware [None req-0c6bc062-cb90-4d50-aabf-8c0503771f7e tempest-InstanceActionsNegativeTestJSON-1719444358 tempest-InstanceActionsNegativeTestJSON-1719444358-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 717.175113] env[63088]: DEBUG nova.virt.hardware [None req-0c6bc062-cb90-4d50-aabf-8c0503771f7e tempest-InstanceActionsNegativeTestJSON-1719444358 tempest-InstanceActionsNegativeTestJSON-1719444358-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 717.176083] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f142367f-587e-4ea4-be65-f96117fcbc04 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.184592] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76a8c39e-db76-483f-ab06-63ace8fdf051 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.200659] env[63088]: ERROR nova.compute.manager [None req-0c6bc062-cb90-4d50-aabf-8c0503771f7e tempest-InstanceActionsNegativeTestJSON-1719444358 tempest-InstanceActionsNegativeTestJSON-1719444358-project-member] [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9c453a07-3eda-4075-81ce-a4516dc00de7, please check neutron logs for more information. [ 717.200659] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] Traceback (most recent call last): [ 717.200659] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 717.200659] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] yield resources [ 717.200659] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 717.200659] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] self.driver.spawn(context, instance, image_meta, [ 717.200659] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 717.200659] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] self._vmops.spawn(context, instance, image_meta, injected_files, [ 717.200659] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 717.200659] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] vm_ref = self.build_virtual_machine(instance, [ 717.200659] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 717.201022] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] vif_infos = vmwarevif.get_vif_info(self._session, [ 717.201022] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 717.201022] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] for vif in network_info: [ 717.201022] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 717.201022] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] return self._sync_wrapper(fn, *args, **kwargs) [ 717.201022] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 717.201022] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] self.wait() [ 717.201022] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 717.201022] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] self[:] = self._gt.wait() [ 717.201022] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 717.201022] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] return self._exit_event.wait() [ 717.201022] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 717.201022] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] current.throw(*self._exc) [ 717.201366] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 717.201366] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] result = function(*args, **kwargs) [ 717.201366] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 717.201366] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] return func(*args, **kwargs) [ 717.201366] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 717.201366] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] raise e [ 717.201366] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 717.201366] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] nwinfo = self.network_api.allocate_for_instance( [ 717.201366] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 717.201366] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] created_port_ids = self._update_ports_for_instance( [ 717.201366] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 717.201366] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] with excutils.save_and_reraise_exception(): [ 717.201366] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 717.201692] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] self.force_reraise() [ 717.201692] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 717.201692] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] raise self.value [ 717.201692] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 717.201692] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] updated_port = self._update_port( [ 717.201692] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 717.201692] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] _ensure_no_port_binding_failure(port) [ 717.201692] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 717.201692] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] raise exception.PortBindingFailed(port_id=port['id']) [ 717.201692] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] nova.exception.PortBindingFailed: Binding failed for port 9c453a07-3eda-4075-81ce-a4516dc00de7, please check neutron logs for more information. [ 717.201692] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] [ 717.201692] env[63088]: INFO nova.compute.manager [None req-0c6bc062-cb90-4d50-aabf-8c0503771f7e tempest-InstanceActionsNegativeTestJSON-1719444358 tempest-InstanceActionsNegativeTestJSON-1719444358-project-member] [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] Terminating instance [ 717.202954] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0c6bc062-cb90-4d50-aabf-8c0503771f7e tempest-InstanceActionsNegativeTestJSON-1719444358 tempest-InstanceActionsNegativeTestJSON-1719444358-project-member] Acquiring lock "refresh_cache-20611ed4-97b3-4d6d-bef7-c13e1235c401" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 717.203124] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0c6bc062-cb90-4d50-aabf-8c0503771f7e tempest-InstanceActionsNegativeTestJSON-1719444358 tempest-InstanceActionsNegativeTestJSON-1719444358-project-member] Acquired lock "refresh_cache-20611ed4-97b3-4d6d-bef7-c13e1235c401" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 717.203295] env[63088]: DEBUG nova.network.neutron [None req-0c6bc062-cb90-4d50-aabf-8c0503771f7e tempest-InstanceActionsNegativeTestJSON-1719444358 tempest-InstanceActionsNegativeTestJSON-1719444358-project-member] [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 717.302883] env[63088]: DEBUG oslo_concurrency.lockutils [req-5298851f-a6eb-480d-986a-2b5df18bdab2 req-9f23c420-e345-407d-a028-ee9861cffdad service nova] Releasing lock "refresh_cache-1219fe86-f218-4a2b-84ba-4a377829386d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 717.303164] env[63088]: DEBUG nova.compute.manager [req-5298851f-a6eb-480d-986a-2b5df18bdab2 req-9f23c420-e345-407d-a028-ee9861cffdad service nova] [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] Received event network-vif-deleted-c47ac4c5-a4a8-4766-bf26-05d24b003141 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 717.573561] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a2a0ac90-0e7a-45d1-a070-f7ee6be01d68 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.453s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 717.574081] env[63088]: DEBUG nova.compute.manager [None req-a2a0ac90-0e7a-45d1-a070-f7ee6be01d68 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 717.578089] env[63088]: DEBUG oslo_concurrency.lockutils [None req-ae448e13-3c80-41d5-9b1f-e650107512d4 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.207s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 717.579465] env[63088]: INFO nova.compute.claims [None req-ae448e13-3c80-41d5-9b1f-e650107512d4 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 717.723683] env[63088]: DEBUG nova.network.neutron [None req-0c6bc062-cb90-4d50-aabf-8c0503771f7e tempest-InstanceActionsNegativeTestJSON-1719444358 tempest-InstanceActionsNegativeTestJSON-1719444358-project-member] [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 717.814570] env[63088]: DEBUG nova.network.neutron [None req-0c6bc062-cb90-4d50-aabf-8c0503771f7e tempest-InstanceActionsNegativeTestJSON-1719444358 tempest-InstanceActionsNegativeTestJSON-1719444358-project-member] [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 718.088875] env[63088]: DEBUG nova.compute.utils [None req-a2a0ac90-0e7a-45d1-a070-f7ee6be01d68 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 718.091461] env[63088]: DEBUG nova.compute.manager [None req-a2a0ac90-0e7a-45d1-a070-f7ee6be01d68 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 718.091628] env[63088]: DEBUG nova.network.neutron [None req-a2a0ac90-0e7a-45d1-a070-f7ee6be01d68 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 718.139053] env[63088]: DEBUG nova.policy [None req-a2a0ac90-0e7a-45d1-a070-f7ee6be01d68 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4424adc5bbe34a91b669c3a907882644', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f33f2701fad94864a8c406a404bc0a42', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 718.212981] env[63088]: DEBUG nova.compute.manager [req-65207dbd-976c-436a-b607-0f228afaa14b req-9a51149b-b5d8-4c36-b99b-8f7e521b6b86 service nova] [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] Received event network-changed-9c453a07-3eda-4075-81ce-a4516dc00de7 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 718.212981] env[63088]: DEBUG nova.compute.manager [req-65207dbd-976c-436a-b607-0f228afaa14b req-9a51149b-b5d8-4c36-b99b-8f7e521b6b86 service nova] [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] Refreshing instance network info cache due to event network-changed-9c453a07-3eda-4075-81ce-a4516dc00de7. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 718.212981] env[63088]: DEBUG oslo_concurrency.lockutils [req-65207dbd-976c-436a-b607-0f228afaa14b req-9a51149b-b5d8-4c36-b99b-8f7e521b6b86 service nova] Acquiring lock "refresh_cache-20611ed4-97b3-4d6d-bef7-c13e1235c401" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 718.318186] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0c6bc062-cb90-4d50-aabf-8c0503771f7e tempest-InstanceActionsNegativeTestJSON-1719444358 tempest-InstanceActionsNegativeTestJSON-1719444358-project-member] Releasing lock "refresh_cache-20611ed4-97b3-4d6d-bef7-c13e1235c401" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 718.318186] env[63088]: DEBUG nova.compute.manager [None req-0c6bc062-cb90-4d50-aabf-8c0503771f7e tempest-InstanceActionsNegativeTestJSON-1719444358 tempest-InstanceActionsNegativeTestJSON-1719444358-project-member] [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 718.318340] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-0c6bc062-cb90-4d50-aabf-8c0503771f7e tempest-InstanceActionsNegativeTestJSON-1719444358 tempest-InstanceActionsNegativeTestJSON-1719444358-project-member] [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 718.319180] env[63088]: DEBUG oslo_concurrency.lockutils [req-65207dbd-976c-436a-b607-0f228afaa14b req-9a51149b-b5d8-4c36-b99b-8f7e521b6b86 service nova] Acquired lock "refresh_cache-20611ed4-97b3-4d6d-bef7-c13e1235c401" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 718.319180] env[63088]: DEBUG nova.network.neutron [req-65207dbd-976c-436a-b607-0f228afaa14b req-9a51149b-b5d8-4c36-b99b-8f7e521b6b86 service nova] [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] Refreshing network info cache for port 9c453a07-3eda-4075-81ce-a4516dc00de7 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 718.320605] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-89042d00-145d-4318-94fb-e8d7f3b1264b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.334231] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-499b1882-df4a-49c0-8eec-d9d30873fc12 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.356692] env[63088]: WARNING nova.virt.vmwareapi.vmops [None req-0c6bc062-cb90-4d50-aabf-8c0503771f7e tempest-InstanceActionsNegativeTestJSON-1719444358 tempest-InstanceActionsNegativeTestJSON-1719444358-project-member] [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 20611ed4-97b3-4d6d-bef7-c13e1235c401 could not be found. [ 718.356923] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-0c6bc062-cb90-4d50-aabf-8c0503771f7e tempest-InstanceActionsNegativeTestJSON-1719444358 tempest-InstanceActionsNegativeTestJSON-1719444358-project-member] [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 718.357192] env[63088]: INFO nova.compute.manager [None req-0c6bc062-cb90-4d50-aabf-8c0503771f7e tempest-InstanceActionsNegativeTestJSON-1719444358 tempest-InstanceActionsNegativeTestJSON-1719444358-project-member] [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] Took 0.04 seconds to destroy the instance on the hypervisor. [ 718.357522] env[63088]: DEBUG oslo.service.loopingcall [None req-0c6bc062-cb90-4d50-aabf-8c0503771f7e tempest-InstanceActionsNegativeTestJSON-1719444358 tempest-InstanceActionsNegativeTestJSON-1719444358-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 718.357751] env[63088]: DEBUG nova.compute.manager [-] [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 718.357842] env[63088]: DEBUG nova.network.neutron [-] [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 718.378203] env[63088]: DEBUG nova.network.neutron [-] [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 718.398319] env[63088]: DEBUG nova.network.neutron [None req-a2a0ac90-0e7a-45d1-a070-f7ee6be01d68 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] Successfully created port: 69404326-63f0-4b81-8180-eae329f85f93 {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 718.592547] env[63088]: DEBUG nova.compute.manager [None req-a2a0ac90-0e7a-45d1-a070-f7ee6be01d68 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 718.851504] env[63088]: DEBUG nova.network.neutron [req-65207dbd-976c-436a-b607-0f228afaa14b req-9a51149b-b5d8-4c36-b99b-8f7e521b6b86 service nova] [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 718.880606] env[63088]: DEBUG nova.network.neutron [-] [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 718.942639] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fff9e975-54cc-476e-bf21-0d79f101bf5c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.952464] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f081b2e7-7baa-48e3-bc83-7090197b21c1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.984780] env[63088]: DEBUG nova.network.neutron [req-65207dbd-976c-436a-b607-0f228afaa14b req-9a51149b-b5d8-4c36-b99b-8f7e521b6b86 service nova] [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 718.986617] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abe1f611-f284-4fda-ac6c-1abb009034c8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.998111] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd97c3dd-4704-4d08-8602-bd2be78c800f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.012320] env[63088]: DEBUG nova.compute.provider_tree [None req-ae448e13-3c80-41d5-9b1f-e650107512d4 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 719.384024] env[63088]: INFO nova.compute.manager [-] [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] Took 1.03 seconds to deallocate network for instance. [ 719.386878] env[63088]: DEBUG nova.compute.claims [None req-0c6bc062-cb90-4d50-aabf-8c0503771f7e tempest-InstanceActionsNegativeTestJSON-1719444358 tempest-InstanceActionsNegativeTestJSON-1719444358-project-member] [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] Aborting claim: {{(pid=63088) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 719.387098] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0c6bc062-cb90-4d50-aabf-8c0503771f7e tempest-InstanceActionsNegativeTestJSON-1719444358 tempest-InstanceActionsNegativeTestJSON-1719444358-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 719.497233] env[63088]: DEBUG oslo_concurrency.lockutils [req-65207dbd-976c-436a-b607-0f228afaa14b req-9a51149b-b5d8-4c36-b99b-8f7e521b6b86 service nova] Releasing lock "refresh_cache-20611ed4-97b3-4d6d-bef7-c13e1235c401" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 719.497233] env[63088]: DEBUG nova.compute.manager [req-65207dbd-976c-436a-b607-0f228afaa14b req-9a51149b-b5d8-4c36-b99b-8f7e521b6b86 service nova] [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] Received event network-vif-deleted-9c453a07-3eda-4075-81ce-a4516dc00de7 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 719.516132] env[63088]: DEBUG nova.scheduler.client.report [None req-ae448e13-3c80-41d5-9b1f-e650107512d4 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 719.545059] env[63088]: ERROR nova.compute.manager [None req-a2a0ac90-0e7a-45d1-a070-f7ee6be01d68 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 69404326-63f0-4b81-8180-eae329f85f93, please check neutron logs for more information. [ 719.545059] env[63088]: ERROR nova.compute.manager Traceback (most recent call last): [ 719.545059] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 719.545059] env[63088]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 719.545059] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 719.545059] env[63088]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 719.545059] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 719.545059] env[63088]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 719.545059] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 719.545059] env[63088]: ERROR nova.compute.manager self.force_reraise() [ 719.545059] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 719.545059] env[63088]: ERROR nova.compute.manager raise self.value [ 719.545059] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 719.545059] env[63088]: ERROR nova.compute.manager updated_port = self._update_port( [ 719.545059] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 719.545059] env[63088]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 719.545546] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 719.545546] env[63088]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 719.545546] env[63088]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 69404326-63f0-4b81-8180-eae329f85f93, please check neutron logs for more information. [ 719.545546] env[63088]: ERROR nova.compute.manager [ 719.545546] env[63088]: Traceback (most recent call last): [ 719.545546] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 719.545546] env[63088]: listener.cb(fileno) [ 719.545546] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 719.545546] env[63088]: result = function(*args, **kwargs) [ 719.545546] env[63088]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 719.545546] env[63088]: return func(*args, **kwargs) [ 719.545546] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 719.545546] env[63088]: raise e [ 719.545546] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 719.545546] env[63088]: nwinfo = self.network_api.allocate_for_instance( [ 719.545546] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 719.545546] env[63088]: created_port_ids = self._update_ports_for_instance( [ 719.545546] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 719.545546] env[63088]: with excutils.save_and_reraise_exception(): [ 719.545546] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 719.545546] env[63088]: self.force_reraise() [ 719.545546] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 719.545546] env[63088]: raise self.value [ 719.545546] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 719.545546] env[63088]: updated_port = self._update_port( [ 719.545546] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 719.545546] env[63088]: _ensure_no_port_binding_failure(port) [ 719.545546] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 719.545546] env[63088]: raise exception.PortBindingFailed(port_id=port['id']) [ 719.546320] env[63088]: nova.exception.PortBindingFailed: Binding failed for port 69404326-63f0-4b81-8180-eae329f85f93, please check neutron logs for more information. [ 719.546320] env[63088]: Removing descriptor: 17 [ 719.606439] env[63088]: DEBUG nova.compute.manager [None req-a2a0ac90-0e7a-45d1-a070-f7ee6be01d68 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 719.631663] env[63088]: DEBUG nova.virt.hardware [None req-a2a0ac90-0e7a-45d1-a070-f7ee6be01d68 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 719.631907] env[63088]: DEBUG nova.virt.hardware [None req-a2a0ac90-0e7a-45d1-a070-f7ee6be01d68 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 719.632079] env[63088]: DEBUG nova.virt.hardware [None req-a2a0ac90-0e7a-45d1-a070-f7ee6be01d68 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 719.632272] env[63088]: DEBUG nova.virt.hardware [None req-a2a0ac90-0e7a-45d1-a070-f7ee6be01d68 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 719.632416] env[63088]: DEBUG nova.virt.hardware [None req-a2a0ac90-0e7a-45d1-a070-f7ee6be01d68 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 719.632558] env[63088]: DEBUG nova.virt.hardware [None req-a2a0ac90-0e7a-45d1-a070-f7ee6be01d68 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 719.632758] env[63088]: DEBUG nova.virt.hardware [None req-a2a0ac90-0e7a-45d1-a070-f7ee6be01d68 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 719.632934] env[63088]: DEBUG nova.virt.hardware [None req-a2a0ac90-0e7a-45d1-a070-f7ee6be01d68 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 719.633089] env[63088]: DEBUG nova.virt.hardware [None req-a2a0ac90-0e7a-45d1-a070-f7ee6be01d68 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 719.633258] env[63088]: DEBUG nova.virt.hardware [None req-a2a0ac90-0e7a-45d1-a070-f7ee6be01d68 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 719.633429] env[63088]: DEBUG nova.virt.hardware [None req-a2a0ac90-0e7a-45d1-a070-f7ee6be01d68 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 719.634302] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11e62b9a-8a1f-48af-9487-edb9a4f8da0a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.642022] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e050664-331b-4699-83bd-89ddbe82551c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.655235] env[63088]: ERROR nova.compute.manager [None req-a2a0ac90-0e7a-45d1-a070-f7ee6be01d68 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 69404326-63f0-4b81-8180-eae329f85f93, please check neutron logs for more information. [ 719.655235] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] Traceback (most recent call last): [ 719.655235] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 719.655235] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] yield resources [ 719.655235] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 719.655235] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] self.driver.spawn(context, instance, image_meta, [ 719.655235] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 719.655235] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 719.655235] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 719.655235] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] vm_ref = self.build_virtual_machine(instance, [ 719.655235] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 719.655627] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] vif_infos = vmwarevif.get_vif_info(self._session, [ 719.655627] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 719.655627] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] for vif in network_info: [ 719.655627] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 719.655627] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] return self._sync_wrapper(fn, *args, **kwargs) [ 719.655627] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 719.655627] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] self.wait() [ 719.655627] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 719.655627] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] self[:] = self._gt.wait() [ 719.655627] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 719.655627] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] return self._exit_event.wait() [ 719.655627] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 719.655627] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] current.throw(*self._exc) [ 719.655956] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 719.655956] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] result = function(*args, **kwargs) [ 719.655956] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 719.655956] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] return func(*args, **kwargs) [ 719.655956] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 719.655956] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] raise e [ 719.655956] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 719.655956] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] nwinfo = self.network_api.allocate_for_instance( [ 719.655956] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 719.655956] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] created_port_ids = self._update_ports_for_instance( [ 719.655956] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 719.655956] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] with excutils.save_and_reraise_exception(): [ 719.655956] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 719.656311] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] self.force_reraise() [ 719.656311] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 719.656311] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] raise self.value [ 719.656311] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 719.656311] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] updated_port = self._update_port( [ 719.656311] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 719.656311] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] _ensure_no_port_binding_failure(port) [ 719.656311] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 719.656311] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] raise exception.PortBindingFailed(port_id=port['id']) [ 719.656311] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] nova.exception.PortBindingFailed: Binding failed for port 69404326-63f0-4b81-8180-eae329f85f93, please check neutron logs for more information. [ 719.656311] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] [ 719.656311] env[63088]: INFO nova.compute.manager [None req-a2a0ac90-0e7a-45d1-a070-f7ee6be01d68 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] Terminating instance [ 719.657451] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a2a0ac90-0e7a-45d1-a070-f7ee6be01d68 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquiring lock "refresh_cache-eb42c667-b0ae-4807-8f69-88dd671aa7d7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 719.657607] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a2a0ac90-0e7a-45d1-a070-f7ee6be01d68 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquired lock "refresh_cache-eb42c667-b0ae-4807-8f69-88dd671aa7d7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 719.657769] env[63088]: DEBUG nova.network.neutron [None req-a2a0ac90-0e7a-45d1-a070-f7ee6be01d68 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 720.021008] env[63088]: DEBUG oslo_concurrency.lockutils [None req-ae448e13-3c80-41d5-9b1f-e650107512d4 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.443s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 720.021635] env[63088]: DEBUG nova.compute.manager [None req-ae448e13-3c80-41d5-9b1f-e650107512d4 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 720.026115] env[63088]: DEBUG oslo_concurrency.lockutils [None req-7af44f64-0067-49a1-ac0c-3ee5b8047867 tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.773s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 720.027514] env[63088]: INFO nova.compute.claims [None req-7af44f64-0067-49a1-ac0c-3ee5b8047867 tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 720.178498] env[63088]: DEBUG nova.network.neutron [None req-a2a0ac90-0e7a-45d1-a070-f7ee6be01d68 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 720.239807] env[63088]: DEBUG nova.compute.manager [req-3eac5331-927c-4800-9365-e6a217ac7fe3 req-d596f11c-874e-4927-9c90-7b767f3bd3a8 service nova] [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] Received event network-changed-69404326-63f0-4b81-8180-eae329f85f93 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 720.240008] env[63088]: DEBUG nova.compute.manager [req-3eac5331-927c-4800-9365-e6a217ac7fe3 req-d596f11c-874e-4927-9c90-7b767f3bd3a8 service nova] [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] Refreshing instance network info cache due to event network-changed-69404326-63f0-4b81-8180-eae329f85f93. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 720.240203] env[63088]: DEBUG oslo_concurrency.lockutils [req-3eac5331-927c-4800-9365-e6a217ac7fe3 req-d596f11c-874e-4927-9c90-7b767f3bd3a8 service nova] Acquiring lock "refresh_cache-eb42c667-b0ae-4807-8f69-88dd671aa7d7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 720.267323] env[63088]: DEBUG nova.network.neutron [None req-a2a0ac90-0e7a-45d1-a070-f7ee6be01d68 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 720.532616] env[63088]: DEBUG nova.compute.utils [None req-ae448e13-3c80-41d5-9b1f-e650107512d4 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 720.537337] env[63088]: DEBUG nova.compute.manager [None req-ae448e13-3c80-41d5-9b1f-e650107512d4 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 720.537337] env[63088]: DEBUG nova.network.neutron [None req-ae448e13-3c80-41d5-9b1f-e650107512d4 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 720.578970] env[63088]: DEBUG nova.policy [None req-ae448e13-3c80-41d5-9b1f-e650107512d4 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '32d5ff9b3ea345e4ac56010aa917e51b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '03e20d606e654362acbe2b36fe499ae3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 720.769498] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a2a0ac90-0e7a-45d1-a070-f7ee6be01d68 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Releasing lock "refresh_cache-eb42c667-b0ae-4807-8f69-88dd671aa7d7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 720.769921] env[63088]: DEBUG nova.compute.manager [None req-a2a0ac90-0e7a-45d1-a070-f7ee6be01d68 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 720.770241] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-a2a0ac90-0e7a-45d1-a070-f7ee6be01d68 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 720.770561] env[63088]: DEBUG oslo_concurrency.lockutils [req-3eac5331-927c-4800-9365-e6a217ac7fe3 req-d596f11c-874e-4927-9c90-7b767f3bd3a8 service nova] Acquired lock "refresh_cache-eb42c667-b0ae-4807-8f69-88dd671aa7d7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 720.770727] env[63088]: DEBUG nova.network.neutron [req-3eac5331-927c-4800-9365-e6a217ac7fe3 req-d596f11c-874e-4927-9c90-7b767f3bd3a8 service nova] [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] Refreshing network info cache for port 69404326-63f0-4b81-8180-eae329f85f93 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 720.771756] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c8fbbca6-4238-4d11-bcc1-40e31da6f110 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.783292] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f8da6e2-f2ed-4524-8163-6c7c720b8e0a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.803510] env[63088]: WARNING nova.virt.vmwareapi.vmops [None req-a2a0ac90-0e7a-45d1-a070-f7ee6be01d68 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance eb42c667-b0ae-4807-8f69-88dd671aa7d7 could not be found. [ 720.803510] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-a2a0ac90-0e7a-45d1-a070-f7ee6be01d68 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 720.803510] env[63088]: INFO nova.compute.manager [None req-a2a0ac90-0e7a-45d1-a070-f7ee6be01d68 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] Took 0.03 seconds to destroy the instance on the hypervisor. [ 720.803678] env[63088]: DEBUG oslo.service.loopingcall [None req-a2a0ac90-0e7a-45d1-a070-f7ee6be01d68 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 720.803805] env[63088]: DEBUG nova.compute.manager [-] [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 720.803861] env[63088]: DEBUG nova.network.neutron [-] [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 720.821110] env[63088]: DEBUG nova.network.neutron [-] [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 720.877728] env[63088]: DEBUG nova.network.neutron [None req-ae448e13-3c80-41d5-9b1f-e650107512d4 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] Successfully created port: c7c50755-fc74-476d-8236-768894ed1858 {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 721.040430] env[63088]: DEBUG nova.compute.manager [None req-ae448e13-3c80-41d5-9b1f-e650107512d4 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 721.309506] env[63088]: DEBUG nova.network.neutron [req-3eac5331-927c-4800-9365-e6a217ac7fe3 req-d596f11c-874e-4927-9c90-7b767f3bd3a8 service nova] [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 721.322798] env[63088]: DEBUG nova.network.neutron [-] [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 721.380285] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1924f95b-ff4b-4dba-8075-959e6248885e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.387877] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-012cadb3-028d-4cdb-baee-710a3be7c0a0 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.392557] env[63088]: DEBUG nova.network.neutron [req-3eac5331-927c-4800-9365-e6a217ac7fe3 req-d596f11c-874e-4927-9c90-7b767f3bd3a8 service nova] [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 721.424301] env[63088]: DEBUG oslo_concurrency.lockutils [req-3eac5331-927c-4800-9365-e6a217ac7fe3 req-d596f11c-874e-4927-9c90-7b767f3bd3a8 service nova] Releasing lock "refresh_cache-eb42c667-b0ae-4807-8f69-88dd671aa7d7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 721.424555] env[63088]: DEBUG nova.compute.manager [req-3eac5331-927c-4800-9365-e6a217ac7fe3 req-d596f11c-874e-4927-9c90-7b767f3bd3a8 service nova] [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] Received event network-vif-deleted-69404326-63f0-4b81-8180-eae329f85f93 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 721.425377] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af317e6c-de22-4c47-a8b3-f55d13180c53 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.433332] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d2b43fa-2338-412e-85f2-6279ba600b09 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.447308] env[63088]: DEBUG nova.compute.provider_tree [None req-7af44f64-0067-49a1-ac0c-3ee5b8047867 tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 721.826698] env[63088]: INFO nova.compute.manager [-] [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] Took 1.02 seconds to deallocate network for instance. [ 721.830290] env[63088]: DEBUG nova.compute.claims [None req-a2a0ac90-0e7a-45d1-a070-f7ee6be01d68 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] Aborting claim: {{(pid=63088) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 721.830290] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a2a0ac90-0e7a-45d1-a070-f7ee6be01d68 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 721.834907] env[63088]: ERROR nova.compute.manager [None req-ae448e13-3c80-41d5-9b1f-e650107512d4 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c7c50755-fc74-476d-8236-768894ed1858, please check neutron logs for more information. [ 721.834907] env[63088]: ERROR nova.compute.manager Traceback (most recent call last): [ 721.834907] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 721.834907] env[63088]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 721.834907] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 721.834907] env[63088]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 721.834907] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 721.834907] env[63088]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 721.834907] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 721.834907] env[63088]: ERROR nova.compute.manager self.force_reraise() [ 721.834907] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 721.834907] env[63088]: ERROR nova.compute.manager raise self.value [ 721.834907] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 721.834907] env[63088]: ERROR nova.compute.manager updated_port = self._update_port( [ 721.834907] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 721.834907] env[63088]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 721.835646] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 721.835646] env[63088]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 721.835646] env[63088]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c7c50755-fc74-476d-8236-768894ed1858, please check neutron logs for more information. [ 721.835646] env[63088]: ERROR nova.compute.manager [ 721.835646] env[63088]: Traceback (most recent call last): [ 721.835646] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 721.835646] env[63088]: listener.cb(fileno) [ 721.835646] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 721.835646] env[63088]: result = function(*args, **kwargs) [ 721.835646] env[63088]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 721.835646] env[63088]: return func(*args, **kwargs) [ 721.835646] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 721.835646] env[63088]: raise e [ 721.835646] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 721.835646] env[63088]: nwinfo = self.network_api.allocate_for_instance( [ 721.835646] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 721.835646] env[63088]: created_port_ids = self._update_ports_for_instance( [ 721.835646] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 721.835646] env[63088]: with excutils.save_and_reraise_exception(): [ 721.835646] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 721.835646] env[63088]: self.force_reraise() [ 721.835646] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 721.835646] env[63088]: raise self.value [ 721.835646] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 721.835646] env[63088]: updated_port = self._update_port( [ 721.835646] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 721.835646] env[63088]: _ensure_no_port_binding_failure(port) [ 721.835646] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 721.835646] env[63088]: raise exception.PortBindingFailed(port_id=port['id']) [ 721.836758] env[63088]: nova.exception.PortBindingFailed: Binding failed for port c7c50755-fc74-476d-8236-768894ed1858, please check neutron logs for more information. [ 721.836758] env[63088]: Removing descriptor: 17 [ 721.950105] env[63088]: DEBUG nova.scheduler.client.report [None req-7af44f64-0067-49a1-ac0c-3ee5b8047867 tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 722.053653] env[63088]: DEBUG nova.compute.manager [None req-ae448e13-3c80-41d5-9b1f-e650107512d4 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 722.077873] env[63088]: DEBUG nova.virt.hardware [None req-ae448e13-3c80-41d5-9b1f-e650107512d4 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 722.078137] env[63088]: DEBUG nova.virt.hardware [None req-ae448e13-3c80-41d5-9b1f-e650107512d4 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 722.078292] env[63088]: DEBUG nova.virt.hardware [None req-ae448e13-3c80-41d5-9b1f-e650107512d4 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 722.078527] env[63088]: DEBUG nova.virt.hardware [None req-ae448e13-3c80-41d5-9b1f-e650107512d4 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 722.078606] env[63088]: DEBUG nova.virt.hardware [None req-ae448e13-3c80-41d5-9b1f-e650107512d4 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 722.078762] env[63088]: DEBUG nova.virt.hardware [None req-ae448e13-3c80-41d5-9b1f-e650107512d4 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 722.078968] env[63088]: DEBUG nova.virt.hardware [None req-ae448e13-3c80-41d5-9b1f-e650107512d4 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 722.079165] env[63088]: DEBUG nova.virt.hardware [None req-ae448e13-3c80-41d5-9b1f-e650107512d4 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 722.079338] env[63088]: DEBUG nova.virt.hardware [None req-ae448e13-3c80-41d5-9b1f-e650107512d4 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 722.079496] env[63088]: DEBUG nova.virt.hardware [None req-ae448e13-3c80-41d5-9b1f-e650107512d4 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 722.079663] env[63088]: DEBUG nova.virt.hardware [None req-ae448e13-3c80-41d5-9b1f-e650107512d4 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 722.080546] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-135465d0-dbbb-4fc7-ba00-74b85a58a2ac {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.088430] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f7b6612-1e74-4db0-b28f-8f4da7e7a304 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.101966] env[63088]: ERROR nova.compute.manager [None req-ae448e13-3c80-41d5-9b1f-e650107512d4 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c7c50755-fc74-476d-8236-768894ed1858, please check neutron logs for more information. [ 722.101966] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] Traceback (most recent call last): [ 722.101966] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 722.101966] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] yield resources [ 722.101966] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 722.101966] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] self.driver.spawn(context, instance, image_meta, [ 722.101966] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 722.101966] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] self._vmops.spawn(context, instance, image_meta, injected_files, [ 722.101966] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 722.101966] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] vm_ref = self.build_virtual_machine(instance, [ 722.101966] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 722.102381] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] vif_infos = vmwarevif.get_vif_info(self._session, [ 722.102381] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 722.102381] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] for vif in network_info: [ 722.102381] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 722.102381] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] return self._sync_wrapper(fn, *args, **kwargs) [ 722.102381] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 722.102381] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] self.wait() [ 722.102381] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 722.102381] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] self[:] = self._gt.wait() [ 722.102381] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 722.102381] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] return self._exit_event.wait() [ 722.102381] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 722.102381] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] current.throw(*self._exc) [ 722.102763] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 722.102763] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] result = function(*args, **kwargs) [ 722.102763] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 722.102763] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] return func(*args, **kwargs) [ 722.102763] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 722.102763] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] raise e [ 722.102763] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 722.102763] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] nwinfo = self.network_api.allocate_for_instance( [ 722.102763] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 722.102763] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] created_port_ids = self._update_ports_for_instance( [ 722.102763] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 722.102763] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] with excutils.save_and_reraise_exception(): [ 722.102763] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 722.103156] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] self.force_reraise() [ 722.103156] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 722.103156] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] raise self.value [ 722.103156] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 722.103156] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] updated_port = self._update_port( [ 722.103156] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 722.103156] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] _ensure_no_port_binding_failure(port) [ 722.103156] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 722.103156] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] raise exception.PortBindingFailed(port_id=port['id']) [ 722.103156] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] nova.exception.PortBindingFailed: Binding failed for port c7c50755-fc74-476d-8236-768894ed1858, please check neutron logs for more information. [ 722.103156] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] [ 722.103156] env[63088]: INFO nova.compute.manager [None req-ae448e13-3c80-41d5-9b1f-e650107512d4 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] Terminating instance [ 722.104055] env[63088]: DEBUG oslo_concurrency.lockutils [None req-ae448e13-3c80-41d5-9b1f-e650107512d4 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquiring lock "refresh_cache-2b43620d-94ac-4c56-87df-4066f5d24088" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 722.104214] env[63088]: DEBUG oslo_concurrency.lockutils [None req-ae448e13-3c80-41d5-9b1f-e650107512d4 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquired lock "refresh_cache-2b43620d-94ac-4c56-87df-4066f5d24088" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 722.104375] env[63088]: DEBUG nova.network.neutron [None req-ae448e13-3c80-41d5-9b1f-e650107512d4 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 722.263625] env[63088]: DEBUG nova.compute.manager [req-5e062c46-c409-4ad2-8c48-de4a2b8f9fd3 req-2b6aaece-4025-42e7-912b-9f7ca030feb6 service nova] [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] Received event network-changed-c7c50755-fc74-476d-8236-768894ed1858 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 722.263818] env[63088]: DEBUG nova.compute.manager [req-5e062c46-c409-4ad2-8c48-de4a2b8f9fd3 req-2b6aaece-4025-42e7-912b-9f7ca030feb6 service nova] [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] Refreshing instance network info cache due to event network-changed-c7c50755-fc74-476d-8236-768894ed1858. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 722.264009] env[63088]: DEBUG oslo_concurrency.lockutils [req-5e062c46-c409-4ad2-8c48-de4a2b8f9fd3 req-2b6aaece-4025-42e7-912b-9f7ca030feb6 service nova] Acquiring lock "refresh_cache-2b43620d-94ac-4c56-87df-4066f5d24088" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 722.455260] env[63088]: DEBUG oslo_concurrency.lockutils [None req-7af44f64-0067-49a1-ac0c-3ee5b8047867 tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.429s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 722.455881] env[63088]: DEBUG nova.compute.manager [None req-7af44f64-0067-49a1-ac0c-3ee5b8047867 tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 722.458451] env[63088]: DEBUG oslo_concurrency.lockutils [None req-03263d3e-caae-4f24-95aa-ac7aedd64259 tempest-ImagesOneServerTestJSON-44065314 tempest-ImagesOneServerTestJSON-44065314-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.876s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 722.460353] env[63088]: INFO nova.compute.claims [None req-03263d3e-caae-4f24-95aa-ac7aedd64259 tempest-ImagesOneServerTestJSON-44065314 tempest-ImagesOneServerTestJSON-44065314-project-member] [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 722.623270] env[63088]: DEBUG nova.network.neutron [None req-ae448e13-3c80-41d5-9b1f-e650107512d4 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 722.714673] env[63088]: DEBUG nova.network.neutron [None req-ae448e13-3c80-41d5-9b1f-e650107512d4 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 722.968203] env[63088]: DEBUG nova.compute.utils [None req-7af44f64-0067-49a1-ac0c-3ee5b8047867 tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 722.970094] env[63088]: DEBUG nova.compute.manager [None req-7af44f64-0067-49a1-ac0c-3ee5b8047867 tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 722.970269] env[63088]: DEBUG nova.network.neutron [None req-7af44f64-0067-49a1-ac0c-3ee5b8047867 tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 723.012424] env[63088]: DEBUG nova.policy [None req-7af44f64-0067-49a1-ac0c-3ee5b8047867 tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b9356a34eb6446339640348c78fa5f24', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2805c642e0d245cf92424e6a807c230e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 723.218722] env[63088]: DEBUG oslo_concurrency.lockutils [None req-ae448e13-3c80-41d5-9b1f-e650107512d4 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Releasing lock "refresh_cache-2b43620d-94ac-4c56-87df-4066f5d24088" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 723.219234] env[63088]: DEBUG nova.compute.manager [None req-ae448e13-3c80-41d5-9b1f-e650107512d4 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 723.219430] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-ae448e13-3c80-41d5-9b1f-e650107512d4 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 723.219750] env[63088]: DEBUG oslo_concurrency.lockutils [req-5e062c46-c409-4ad2-8c48-de4a2b8f9fd3 req-2b6aaece-4025-42e7-912b-9f7ca030feb6 service nova] Acquired lock "refresh_cache-2b43620d-94ac-4c56-87df-4066f5d24088" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 723.219915] env[63088]: DEBUG nova.network.neutron [req-5e062c46-c409-4ad2-8c48-de4a2b8f9fd3 req-2b6aaece-4025-42e7-912b-9f7ca030feb6 service nova] [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] Refreshing network info cache for port c7c50755-fc74-476d-8236-768894ed1858 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 723.220973] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fa209c4c-fc2c-47e8-9b38-0a621cadcac0 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.234515] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccd4e7c0-4004-49de-aeca-f7e5585b36e8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.260126] env[63088]: WARNING nova.virt.vmwareapi.vmops [None req-ae448e13-3c80-41d5-9b1f-e650107512d4 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2b43620d-94ac-4c56-87df-4066f5d24088 could not be found. [ 723.260355] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-ae448e13-3c80-41d5-9b1f-e650107512d4 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 723.260535] env[63088]: INFO nova.compute.manager [None req-ae448e13-3c80-41d5-9b1f-e650107512d4 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] Took 0.04 seconds to destroy the instance on the hypervisor. [ 723.260781] env[63088]: DEBUG oslo.service.loopingcall [None req-ae448e13-3c80-41d5-9b1f-e650107512d4 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 723.261015] env[63088]: DEBUG nova.compute.manager [-] [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 723.261116] env[63088]: DEBUG nova.network.neutron [-] [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 723.280244] env[63088]: DEBUG nova.network.neutron [-] [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 723.310038] env[63088]: DEBUG nova.network.neutron [None req-7af44f64-0067-49a1-ac0c-3ee5b8047867 tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] Successfully created port: 047bc58e-8cf5-4c4d-b766-962ea1063144 {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 723.470445] env[63088]: DEBUG nova.compute.manager [None req-7af44f64-0067-49a1-ac0c-3ee5b8047867 tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 723.563870] env[63088]: DEBUG nova.network.neutron [None req-7af44f64-0067-49a1-ac0c-3ee5b8047867 tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] Successfully created port: 6e068d4d-b322-4faa-ab1e-747a2a89a402 {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 723.744286] env[63088]: DEBUG nova.network.neutron [req-5e062c46-c409-4ad2-8c48-de4a2b8f9fd3 req-2b6aaece-4025-42e7-912b-9f7ca030feb6 service nova] [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 723.747615] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87e5214f-95cd-470c-9c6f-6bf21653a05e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.757020] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dacb85a0-aa49-4c19-b0d4-86972b1ff668 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.787993] env[63088]: DEBUG nova.network.neutron [-] [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 723.789958] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1bcef31-4d0b-4579-a786-1cdc4e9fad70 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.798390] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da3d3336-f1de-4e0c-9d21-b13b3f927d34 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.825297] env[63088]: DEBUG nova.compute.provider_tree [None req-03263d3e-caae-4f24-95aa-ac7aedd64259 tempest-ImagesOneServerTestJSON-44065314 tempest-ImagesOneServerTestJSON-44065314-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 723.875938] env[63088]: DEBUG nova.network.neutron [req-5e062c46-c409-4ad2-8c48-de4a2b8f9fd3 req-2b6aaece-4025-42e7-912b-9f7ca030feb6 service nova] [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 724.294344] env[63088]: INFO nova.compute.manager [-] [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] Took 1.03 seconds to deallocate network for instance. [ 724.296864] env[63088]: DEBUG nova.compute.claims [None req-ae448e13-3c80-41d5-9b1f-e650107512d4 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] Aborting claim: {{(pid=63088) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 724.297036] env[63088]: DEBUG oslo_concurrency.lockutils [None req-ae448e13-3c80-41d5-9b1f-e650107512d4 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 724.328636] env[63088]: DEBUG nova.scheduler.client.report [None req-03263d3e-caae-4f24-95aa-ac7aedd64259 tempest-ImagesOneServerTestJSON-44065314 tempest-ImagesOneServerTestJSON-44065314-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 724.375414] env[63088]: DEBUG nova.compute.manager [req-783c708a-85ca-4e1e-a465-5387fab6ba00 req-7c36c459-0ac9-4514-b2e8-7c05b0ee3853 service nova] [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] Received event network-changed-047bc58e-8cf5-4c4d-b766-962ea1063144 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 724.375665] env[63088]: DEBUG nova.compute.manager [req-783c708a-85ca-4e1e-a465-5387fab6ba00 req-7c36c459-0ac9-4514-b2e8-7c05b0ee3853 service nova] [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] Refreshing instance network info cache due to event network-changed-047bc58e-8cf5-4c4d-b766-962ea1063144. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 724.375909] env[63088]: DEBUG oslo_concurrency.lockutils [req-783c708a-85ca-4e1e-a465-5387fab6ba00 req-7c36c459-0ac9-4514-b2e8-7c05b0ee3853 service nova] Acquiring lock "refresh_cache-f2b4485c-ebcb-4063-a25a-e2efe6d1758f" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 724.376092] env[63088]: DEBUG oslo_concurrency.lockutils [req-783c708a-85ca-4e1e-a465-5387fab6ba00 req-7c36c459-0ac9-4514-b2e8-7c05b0ee3853 service nova] Acquired lock "refresh_cache-f2b4485c-ebcb-4063-a25a-e2efe6d1758f" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 724.376297] env[63088]: DEBUG nova.network.neutron [req-783c708a-85ca-4e1e-a465-5387fab6ba00 req-7c36c459-0ac9-4514-b2e8-7c05b0ee3853 service nova] [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] Refreshing network info cache for port 047bc58e-8cf5-4c4d-b766-962ea1063144 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 724.378207] env[63088]: DEBUG oslo_concurrency.lockutils [req-5e062c46-c409-4ad2-8c48-de4a2b8f9fd3 req-2b6aaece-4025-42e7-912b-9f7ca030feb6 service nova] Releasing lock "refresh_cache-2b43620d-94ac-4c56-87df-4066f5d24088" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 724.381240] env[63088]: DEBUG nova.compute.manager [req-5e062c46-c409-4ad2-8c48-de4a2b8f9fd3 req-2b6aaece-4025-42e7-912b-9f7ca030feb6 service nova] [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] Received event network-vif-deleted-c7c50755-fc74-476d-8236-768894ed1858 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 724.481901] env[63088]: DEBUG nova.compute.manager [None req-7af44f64-0067-49a1-ac0c-3ee5b8047867 tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 724.510453] env[63088]: DEBUG nova.virt.hardware [None req-7af44f64-0067-49a1-ac0c-3ee5b8047867 tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 724.510775] env[63088]: DEBUG nova.virt.hardware [None req-7af44f64-0067-49a1-ac0c-3ee5b8047867 tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 724.510836] env[63088]: DEBUG nova.virt.hardware [None req-7af44f64-0067-49a1-ac0c-3ee5b8047867 tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 724.510989] env[63088]: DEBUG nova.virt.hardware [None req-7af44f64-0067-49a1-ac0c-3ee5b8047867 tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 724.511151] env[63088]: DEBUG nova.virt.hardware [None req-7af44f64-0067-49a1-ac0c-3ee5b8047867 tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 724.511866] env[63088]: DEBUG nova.virt.hardware [None req-7af44f64-0067-49a1-ac0c-3ee5b8047867 tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 724.511866] env[63088]: DEBUG nova.virt.hardware [None req-7af44f64-0067-49a1-ac0c-3ee5b8047867 tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 724.511866] env[63088]: DEBUG nova.virt.hardware [None req-7af44f64-0067-49a1-ac0c-3ee5b8047867 tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 724.511866] env[63088]: DEBUG nova.virt.hardware [None req-7af44f64-0067-49a1-ac0c-3ee5b8047867 tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 724.512131] env[63088]: DEBUG nova.virt.hardware [None req-7af44f64-0067-49a1-ac0c-3ee5b8047867 tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 724.512193] env[63088]: DEBUG nova.virt.hardware [None req-7af44f64-0067-49a1-ac0c-3ee5b8047867 tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 724.513474] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-612d2ba9-71db-48bf-8e8d-b774de847008 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.522421] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-978ef1aa-5e7a-4ec0-ad2b-099e5bfadf84 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.536340] env[63088]: ERROR nova.compute.manager [None req-7af44f64-0067-49a1-ac0c-3ee5b8047867 tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 047bc58e-8cf5-4c4d-b766-962ea1063144, please check neutron logs for more information. [ 724.536340] env[63088]: ERROR nova.compute.manager Traceback (most recent call last): [ 724.536340] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 724.536340] env[63088]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 724.536340] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 724.536340] env[63088]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 724.536340] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 724.536340] env[63088]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 724.536340] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 724.536340] env[63088]: ERROR nova.compute.manager self.force_reraise() [ 724.536340] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 724.536340] env[63088]: ERROR nova.compute.manager raise self.value [ 724.536340] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 724.536340] env[63088]: ERROR nova.compute.manager updated_port = self._update_port( [ 724.536340] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 724.536340] env[63088]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 724.536789] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 724.536789] env[63088]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 724.536789] env[63088]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 047bc58e-8cf5-4c4d-b766-962ea1063144, please check neutron logs for more information. [ 724.536789] env[63088]: ERROR nova.compute.manager [ 724.536789] env[63088]: Traceback (most recent call last): [ 724.536789] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 724.536789] env[63088]: listener.cb(fileno) [ 724.536789] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 724.536789] env[63088]: result = function(*args, **kwargs) [ 724.536789] env[63088]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 724.536789] env[63088]: return func(*args, **kwargs) [ 724.536789] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 724.536789] env[63088]: raise e [ 724.536789] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 724.536789] env[63088]: nwinfo = self.network_api.allocate_for_instance( [ 724.536789] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 724.536789] env[63088]: created_port_ids = self._update_ports_for_instance( [ 724.536789] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 724.536789] env[63088]: with excutils.save_and_reraise_exception(): [ 724.536789] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 724.536789] env[63088]: self.force_reraise() [ 724.536789] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 724.536789] env[63088]: raise self.value [ 724.536789] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 724.536789] env[63088]: updated_port = self._update_port( [ 724.536789] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 724.536789] env[63088]: _ensure_no_port_binding_failure(port) [ 724.536789] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 724.536789] env[63088]: raise exception.PortBindingFailed(port_id=port['id']) [ 724.537580] env[63088]: nova.exception.PortBindingFailed: Binding failed for port 047bc58e-8cf5-4c4d-b766-962ea1063144, please check neutron logs for more information. [ 724.537580] env[63088]: Removing descriptor: 17 [ 724.537580] env[63088]: ERROR nova.compute.manager [None req-7af44f64-0067-49a1-ac0c-3ee5b8047867 tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 047bc58e-8cf5-4c4d-b766-962ea1063144, please check neutron logs for more information. [ 724.537580] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] Traceback (most recent call last): [ 724.537580] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 724.537580] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] yield resources [ 724.537580] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 724.537580] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] self.driver.spawn(context, instance, image_meta, [ 724.537580] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 724.537580] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 724.537580] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 724.537580] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] vm_ref = self.build_virtual_machine(instance, [ 724.537957] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 724.537957] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] vif_infos = vmwarevif.get_vif_info(self._session, [ 724.537957] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 724.537957] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] for vif in network_info: [ 724.537957] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 724.537957] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] return self._sync_wrapper(fn, *args, **kwargs) [ 724.537957] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 724.537957] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] self.wait() [ 724.537957] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 724.537957] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] self[:] = self._gt.wait() [ 724.537957] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 724.537957] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] return self._exit_event.wait() [ 724.537957] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 724.538320] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] result = hub.switch() [ 724.538320] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 724.538320] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] return self.greenlet.switch() [ 724.538320] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 724.538320] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] result = function(*args, **kwargs) [ 724.538320] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 724.538320] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] return func(*args, **kwargs) [ 724.538320] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 724.538320] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] raise e [ 724.538320] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 724.538320] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] nwinfo = self.network_api.allocate_for_instance( [ 724.538320] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 724.538320] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] created_port_ids = self._update_ports_for_instance( [ 724.538711] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 724.538711] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] with excutils.save_and_reraise_exception(): [ 724.538711] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 724.538711] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] self.force_reraise() [ 724.538711] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 724.538711] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] raise self.value [ 724.538711] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 724.538711] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] updated_port = self._update_port( [ 724.538711] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 724.538711] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] _ensure_no_port_binding_failure(port) [ 724.538711] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 724.538711] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] raise exception.PortBindingFailed(port_id=port['id']) [ 724.539211] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] nova.exception.PortBindingFailed: Binding failed for port 047bc58e-8cf5-4c4d-b766-962ea1063144, please check neutron logs for more information. [ 724.539211] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] [ 724.539211] env[63088]: INFO nova.compute.manager [None req-7af44f64-0067-49a1-ac0c-3ee5b8047867 tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] Terminating instance [ 724.540091] env[63088]: DEBUG oslo_concurrency.lockutils [None req-7af44f64-0067-49a1-ac0c-3ee5b8047867 tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] Acquiring lock "refresh_cache-f2b4485c-ebcb-4063-a25a-e2efe6d1758f" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 724.833439] env[63088]: DEBUG oslo_concurrency.lockutils [None req-03263d3e-caae-4f24-95aa-ac7aedd64259 tempest-ImagesOneServerTestJSON-44065314 tempest-ImagesOneServerTestJSON-44065314-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.375s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 724.833996] env[63088]: DEBUG nova.compute.manager [None req-03263d3e-caae-4f24-95aa-ac7aedd64259 tempest-ImagesOneServerTestJSON-44065314 tempest-ImagesOneServerTestJSON-44065314-project-member] [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 724.836773] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0b29042d-57c3-47a9-a926-f67d685a4740 tempest-TenantUsagesTestJSON-1054626739 tempest-TenantUsagesTestJSON-1054626739-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.736s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 724.924417] env[63088]: DEBUG nova.network.neutron [req-783c708a-85ca-4e1e-a465-5387fab6ba00 req-7c36c459-0ac9-4514-b2e8-7c05b0ee3853 service nova] [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 725.018833] env[63088]: DEBUG nova.network.neutron [req-783c708a-85ca-4e1e-a465-5387fab6ba00 req-7c36c459-0ac9-4514-b2e8-7c05b0ee3853 service nova] [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 725.340979] env[63088]: DEBUG nova.compute.utils [None req-03263d3e-caae-4f24-95aa-ac7aedd64259 tempest-ImagesOneServerTestJSON-44065314 tempest-ImagesOneServerTestJSON-44065314-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 725.345308] env[63088]: DEBUG nova.compute.manager [None req-03263d3e-caae-4f24-95aa-ac7aedd64259 tempest-ImagesOneServerTestJSON-44065314 tempest-ImagesOneServerTestJSON-44065314-project-member] [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 725.345478] env[63088]: DEBUG nova.network.neutron [None req-03263d3e-caae-4f24-95aa-ac7aedd64259 tempest-ImagesOneServerTestJSON-44065314 tempest-ImagesOneServerTestJSON-44065314-project-member] [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 725.399760] env[63088]: DEBUG nova.policy [None req-03263d3e-caae-4f24-95aa-ac7aedd64259 tempest-ImagesOneServerTestJSON-44065314 tempest-ImagesOneServerTestJSON-44065314-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ac470c1f2b704558bff18e32bd9f2c7f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8fea6b40cf664c2b8fa3739195295937', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 725.522191] env[63088]: DEBUG oslo_concurrency.lockutils [req-783c708a-85ca-4e1e-a465-5387fab6ba00 req-7c36c459-0ac9-4514-b2e8-7c05b0ee3853 service nova] Releasing lock "refresh_cache-f2b4485c-ebcb-4063-a25a-e2efe6d1758f" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 725.525603] env[63088]: DEBUG oslo_concurrency.lockutils [None req-7af44f64-0067-49a1-ac0c-3ee5b8047867 tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] Acquired lock "refresh_cache-f2b4485c-ebcb-4063-a25a-e2efe6d1758f" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 725.525603] env[63088]: DEBUG nova.network.neutron [None req-7af44f64-0067-49a1-ac0c-3ee5b8047867 tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 725.604085] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ca28b31-1de7-41f1-be60-10b1375e7181 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.612756] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ceb11078-351e-4bb6-ba54-7feaf9544d17 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.645684] env[63088]: DEBUG nova.network.neutron [None req-03263d3e-caae-4f24-95aa-ac7aedd64259 tempest-ImagesOneServerTestJSON-44065314 tempest-ImagesOneServerTestJSON-44065314-project-member] [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] Successfully created port: 403c6f43-6162-47c8-a025-754113fc33d6 {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 725.648073] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9725df7-832c-40f7-9b29-037d0cc8ad17 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.656387] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1429abfe-c138-4646-ba7d-40b05542e922 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.672243] env[63088]: DEBUG nova.compute.provider_tree [None req-0b29042d-57c3-47a9-a926-f67d685a4740 tempest-TenantUsagesTestJSON-1054626739 tempest-TenantUsagesTestJSON-1054626739-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 725.847733] env[63088]: DEBUG nova.compute.manager [None req-03263d3e-caae-4f24-95aa-ac7aedd64259 tempest-ImagesOneServerTestJSON-44065314 tempest-ImagesOneServerTestJSON-44065314-project-member] [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 726.074531] env[63088]: DEBUG nova.network.neutron [None req-7af44f64-0067-49a1-ac0c-3ee5b8047867 tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 726.172343] env[63088]: DEBUG nova.scheduler.client.report [None req-0b29042d-57c3-47a9-a926-f67d685a4740 tempest-TenantUsagesTestJSON-1054626739 tempest-TenantUsagesTestJSON-1054626739-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 726.188638] env[63088]: DEBUG nova.network.neutron [None req-7af44f64-0067-49a1-ac0c-3ee5b8047867 tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 726.419109] env[63088]: DEBUG nova.compute.manager [req-877a9f65-60cb-4d69-ae80-48c00de6bf2b req-96c4e89b-ce13-42ab-a63d-fe1f1ac5b46f service nova] [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] Received event network-vif-deleted-047bc58e-8cf5-4c4d-b766-962ea1063144 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 726.420507] env[63088]: DEBUG nova.compute.manager [req-877a9f65-60cb-4d69-ae80-48c00de6bf2b req-96c4e89b-ce13-42ab-a63d-fe1f1ac5b46f service nova] [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] Received event network-changed-403c6f43-6162-47c8-a025-754113fc33d6 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 726.420813] env[63088]: DEBUG nova.compute.manager [req-877a9f65-60cb-4d69-ae80-48c00de6bf2b req-96c4e89b-ce13-42ab-a63d-fe1f1ac5b46f service nova] [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] Refreshing instance network info cache due to event network-changed-403c6f43-6162-47c8-a025-754113fc33d6. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 726.424766] env[63088]: DEBUG oslo_concurrency.lockutils [req-877a9f65-60cb-4d69-ae80-48c00de6bf2b req-96c4e89b-ce13-42ab-a63d-fe1f1ac5b46f service nova] Acquiring lock "refresh_cache-c63c1740-fcec-47fa-a7b9-986b83c4abd9" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 726.425104] env[63088]: DEBUG oslo_concurrency.lockutils [req-877a9f65-60cb-4d69-ae80-48c00de6bf2b req-96c4e89b-ce13-42ab-a63d-fe1f1ac5b46f service nova] Acquired lock "refresh_cache-c63c1740-fcec-47fa-a7b9-986b83c4abd9" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 726.425441] env[63088]: DEBUG nova.network.neutron [req-877a9f65-60cb-4d69-ae80-48c00de6bf2b req-96c4e89b-ce13-42ab-a63d-fe1f1ac5b46f service nova] [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] Refreshing network info cache for port 403c6f43-6162-47c8-a025-754113fc33d6 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 726.565842] env[63088]: ERROR nova.compute.manager [None req-03263d3e-caae-4f24-95aa-ac7aedd64259 tempest-ImagesOneServerTestJSON-44065314 tempest-ImagesOneServerTestJSON-44065314-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 403c6f43-6162-47c8-a025-754113fc33d6, please check neutron logs for more information. [ 726.565842] env[63088]: ERROR nova.compute.manager Traceback (most recent call last): [ 726.565842] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 726.565842] env[63088]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 726.565842] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 726.565842] env[63088]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 726.565842] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 726.565842] env[63088]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 726.565842] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 726.565842] env[63088]: ERROR nova.compute.manager self.force_reraise() [ 726.565842] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 726.565842] env[63088]: ERROR nova.compute.manager raise self.value [ 726.565842] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 726.565842] env[63088]: ERROR nova.compute.manager updated_port = self._update_port( [ 726.565842] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 726.565842] env[63088]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 726.566298] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 726.566298] env[63088]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 726.566298] env[63088]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 403c6f43-6162-47c8-a025-754113fc33d6, please check neutron logs for more information. [ 726.566298] env[63088]: ERROR nova.compute.manager [ 726.566298] env[63088]: Traceback (most recent call last): [ 726.566298] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 726.566298] env[63088]: listener.cb(fileno) [ 726.566298] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 726.566298] env[63088]: result = function(*args, **kwargs) [ 726.566298] env[63088]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 726.566298] env[63088]: return func(*args, **kwargs) [ 726.566298] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 726.566298] env[63088]: raise e [ 726.566298] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 726.566298] env[63088]: nwinfo = self.network_api.allocate_for_instance( [ 726.566298] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 726.566298] env[63088]: created_port_ids = self._update_ports_for_instance( [ 726.566298] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 726.566298] env[63088]: with excutils.save_and_reraise_exception(): [ 726.566298] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 726.566298] env[63088]: self.force_reraise() [ 726.566298] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 726.566298] env[63088]: raise self.value [ 726.566298] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 726.566298] env[63088]: updated_port = self._update_port( [ 726.566298] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 726.566298] env[63088]: _ensure_no_port_binding_failure(port) [ 726.566298] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 726.566298] env[63088]: raise exception.PortBindingFailed(port_id=port['id']) [ 726.567129] env[63088]: nova.exception.PortBindingFailed: Binding failed for port 403c6f43-6162-47c8-a025-754113fc33d6, please check neutron logs for more information. [ 726.567129] env[63088]: Removing descriptor: 17 [ 726.679705] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0b29042d-57c3-47a9-a926-f67d685a4740 tempest-TenantUsagesTestJSON-1054626739 tempest-TenantUsagesTestJSON-1054626739-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.841s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 726.679705] env[63088]: ERROR nova.compute.manager [None req-0b29042d-57c3-47a9-a926-f67d685a4740 tempest-TenantUsagesTestJSON-1054626739 tempest-TenantUsagesTestJSON-1054626739-project-member] [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 32cece70-208a-4d57-bcdf-d63d938be4d3, please check neutron logs for more information. [ 726.679705] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] Traceback (most recent call last): [ 726.679705] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 726.679705] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] self.driver.spawn(context, instance, image_meta, [ 726.679705] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 726.679705] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] self._vmops.spawn(context, instance, image_meta, injected_files, [ 726.679705] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 726.679705] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] vm_ref = self.build_virtual_machine(instance, [ 726.680175] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 726.680175] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] vif_infos = vmwarevif.get_vif_info(self._session, [ 726.680175] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 726.680175] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] for vif in network_info: [ 726.680175] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 726.680175] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] return self._sync_wrapper(fn, *args, **kwargs) [ 726.680175] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 726.680175] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] self.wait() [ 726.680175] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 726.680175] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] self[:] = self._gt.wait() [ 726.680175] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 726.680175] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] return self._exit_event.wait() [ 726.680175] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 726.680616] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] current.throw(*self._exc) [ 726.680616] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 726.680616] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] result = function(*args, **kwargs) [ 726.680616] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 726.680616] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] return func(*args, **kwargs) [ 726.680616] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 726.680616] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] raise e [ 726.680616] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 726.680616] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] nwinfo = self.network_api.allocate_for_instance( [ 726.680616] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 726.680616] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] created_port_ids = self._update_ports_for_instance( [ 726.680616] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 726.680616] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] with excutils.save_and_reraise_exception(): [ 726.680949] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 726.680949] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] self.force_reraise() [ 726.680949] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 726.680949] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] raise self.value [ 726.680949] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 726.680949] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] updated_port = self._update_port( [ 726.680949] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 726.680949] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] _ensure_no_port_binding_failure(port) [ 726.680949] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 726.680949] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] raise exception.PortBindingFailed(port_id=port['id']) [ 726.680949] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] nova.exception.PortBindingFailed: Binding failed for port 32cece70-208a-4d57-bcdf-d63d938be4d3, please check neutron logs for more information. [ 726.680949] env[63088]: ERROR nova.compute.manager [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] [ 726.681311] env[63088]: DEBUG nova.compute.utils [None req-0b29042d-57c3-47a9-a926-f67d685a4740 tempest-TenantUsagesTestJSON-1054626739 tempest-TenantUsagesTestJSON-1054626739-project-member] [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] Binding failed for port 32cece70-208a-4d57-bcdf-d63d938be4d3, please check neutron logs for more information. {{(pid=63088) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 726.681311] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a6077124-74e2-4777-9beb-9998c6275315 tempest-ServersTestBootFromVolume-2105176458 tempest-ServersTestBootFromVolume-2105176458-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.865s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 726.681651] env[63088]: INFO nova.compute.claims [None req-a6077124-74e2-4777-9beb-9998c6275315 tempest-ServersTestBootFromVolume-2105176458 tempest-ServersTestBootFromVolume-2105176458-project-member] [instance: 407db84f-3322-42bc-baac-5528e856af8d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 726.684309] env[63088]: DEBUG nova.compute.manager [None req-0b29042d-57c3-47a9-a926-f67d685a4740 tempest-TenantUsagesTestJSON-1054626739 tempest-TenantUsagesTestJSON-1054626739-project-member] [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] Build of instance 02da88f0-7e3a-4eea-a88d-59a990c69f82 was re-scheduled: Binding failed for port 32cece70-208a-4d57-bcdf-d63d938be4d3, please check neutron logs for more information. {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 726.685026] env[63088]: DEBUG nova.compute.manager [None req-0b29042d-57c3-47a9-a926-f67d685a4740 tempest-TenantUsagesTestJSON-1054626739 tempest-TenantUsagesTestJSON-1054626739-project-member] [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] Unplugging VIFs for instance {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 726.685026] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0b29042d-57c3-47a9-a926-f67d685a4740 tempest-TenantUsagesTestJSON-1054626739 tempest-TenantUsagesTestJSON-1054626739-project-member] Acquiring lock "refresh_cache-02da88f0-7e3a-4eea-a88d-59a990c69f82" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 726.685138] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0b29042d-57c3-47a9-a926-f67d685a4740 tempest-TenantUsagesTestJSON-1054626739 tempest-TenantUsagesTestJSON-1054626739-project-member] Acquired lock "refresh_cache-02da88f0-7e3a-4eea-a88d-59a990c69f82" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 726.685689] env[63088]: DEBUG nova.network.neutron [None req-0b29042d-57c3-47a9-a926-f67d685a4740 tempest-TenantUsagesTestJSON-1054626739 tempest-TenantUsagesTestJSON-1054626739-project-member] [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 726.690703] env[63088]: DEBUG oslo_concurrency.lockutils [None req-7af44f64-0067-49a1-ac0c-3ee5b8047867 tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] Releasing lock "refresh_cache-f2b4485c-ebcb-4063-a25a-e2efe6d1758f" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 726.690911] env[63088]: DEBUG nova.compute.manager [None req-7af44f64-0067-49a1-ac0c-3ee5b8047867 tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 726.691112] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-7af44f64-0067-49a1-ac0c-3ee5b8047867 tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 726.691397] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-95d98898-79d5-404d-99a9-43fe532eab24 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.700414] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6209a35-13a7-43c1-a79b-bd124fc32029 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.722389] env[63088]: WARNING nova.virt.vmwareapi.vmops [None req-7af44f64-0067-49a1-ac0c-3ee5b8047867 tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f2b4485c-ebcb-4063-a25a-e2efe6d1758f could not be found. [ 726.722607] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-7af44f64-0067-49a1-ac0c-3ee5b8047867 tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 726.722781] env[63088]: INFO nova.compute.manager [None req-7af44f64-0067-49a1-ac0c-3ee5b8047867 tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] Took 0.03 seconds to destroy the instance on the hypervisor. [ 726.723027] env[63088]: DEBUG oslo.service.loopingcall [None req-7af44f64-0067-49a1-ac0c-3ee5b8047867 tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 726.723244] env[63088]: DEBUG nova.compute.manager [-] [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 726.723341] env[63088]: DEBUG nova.network.neutron [-] [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 726.749654] env[63088]: DEBUG nova.network.neutron [-] [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 726.855626] env[63088]: DEBUG nova.compute.manager [None req-03263d3e-caae-4f24-95aa-ac7aedd64259 tempest-ImagesOneServerTestJSON-44065314 tempest-ImagesOneServerTestJSON-44065314-project-member] [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 726.880567] env[63088]: DEBUG nova.virt.hardware [None req-03263d3e-caae-4f24-95aa-ac7aedd64259 tempest-ImagesOneServerTestJSON-44065314 tempest-ImagesOneServerTestJSON-44065314-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 726.880809] env[63088]: DEBUG nova.virt.hardware [None req-03263d3e-caae-4f24-95aa-ac7aedd64259 tempest-ImagesOneServerTestJSON-44065314 tempest-ImagesOneServerTestJSON-44065314-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 726.880964] env[63088]: DEBUG nova.virt.hardware [None req-03263d3e-caae-4f24-95aa-ac7aedd64259 tempest-ImagesOneServerTestJSON-44065314 tempest-ImagesOneServerTestJSON-44065314-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 726.881156] env[63088]: DEBUG nova.virt.hardware [None req-03263d3e-caae-4f24-95aa-ac7aedd64259 tempest-ImagesOneServerTestJSON-44065314 tempest-ImagesOneServerTestJSON-44065314-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 726.881302] env[63088]: DEBUG nova.virt.hardware [None req-03263d3e-caae-4f24-95aa-ac7aedd64259 tempest-ImagesOneServerTestJSON-44065314 tempest-ImagesOneServerTestJSON-44065314-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 726.881446] env[63088]: DEBUG nova.virt.hardware [None req-03263d3e-caae-4f24-95aa-ac7aedd64259 tempest-ImagesOneServerTestJSON-44065314 tempest-ImagesOneServerTestJSON-44065314-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 726.881647] env[63088]: DEBUG nova.virt.hardware [None req-03263d3e-caae-4f24-95aa-ac7aedd64259 tempest-ImagesOneServerTestJSON-44065314 tempest-ImagesOneServerTestJSON-44065314-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 726.881804] env[63088]: DEBUG nova.virt.hardware [None req-03263d3e-caae-4f24-95aa-ac7aedd64259 tempest-ImagesOneServerTestJSON-44065314 tempest-ImagesOneServerTestJSON-44065314-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 726.881963] env[63088]: DEBUG nova.virt.hardware [None req-03263d3e-caae-4f24-95aa-ac7aedd64259 tempest-ImagesOneServerTestJSON-44065314 tempest-ImagesOneServerTestJSON-44065314-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 726.882134] env[63088]: DEBUG nova.virt.hardware [None req-03263d3e-caae-4f24-95aa-ac7aedd64259 tempest-ImagesOneServerTestJSON-44065314 tempest-ImagesOneServerTestJSON-44065314-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 726.882308] env[63088]: DEBUG nova.virt.hardware [None req-03263d3e-caae-4f24-95aa-ac7aedd64259 tempest-ImagesOneServerTestJSON-44065314 tempest-ImagesOneServerTestJSON-44065314-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 726.883163] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c7c2e0d-c502-45bb-b729-a8f47b57848d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.890742] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59e33b7f-36be-40b0-a6ce-6f6dee2371c1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.904016] env[63088]: ERROR nova.compute.manager [None req-03263d3e-caae-4f24-95aa-ac7aedd64259 tempest-ImagesOneServerTestJSON-44065314 tempest-ImagesOneServerTestJSON-44065314-project-member] [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 403c6f43-6162-47c8-a025-754113fc33d6, please check neutron logs for more information. [ 726.904016] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] Traceback (most recent call last): [ 726.904016] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 726.904016] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] yield resources [ 726.904016] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 726.904016] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] self.driver.spawn(context, instance, image_meta, [ 726.904016] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 726.904016] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 726.904016] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 726.904016] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] vm_ref = self.build_virtual_machine(instance, [ 726.904016] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 726.904371] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] vif_infos = vmwarevif.get_vif_info(self._session, [ 726.904371] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 726.904371] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] for vif in network_info: [ 726.904371] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 726.904371] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] return self._sync_wrapper(fn, *args, **kwargs) [ 726.904371] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 726.904371] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] self.wait() [ 726.904371] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 726.904371] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] self[:] = self._gt.wait() [ 726.904371] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 726.904371] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] return self._exit_event.wait() [ 726.904371] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 726.904371] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] current.throw(*self._exc) [ 726.904712] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 726.904712] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] result = function(*args, **kwargs) [ 726.904712] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 726.904712] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] return func(*args, **kwargs) [ 726.904712] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 726.904712] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] raise e [ 726.904712] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 726.904712] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] nwinfo = self.network_api.allocate_for_instance( [ 726.904712] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 726.904712] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] created_port_ids = self._update_ports_for_instance( [ 726.904712] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 726.904712] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] with excutils.save_and_reraise_exception(): [ 726.904712] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 726.905126] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] self.force_reraise() [ 726.905126] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 726.905126] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] raise self.value [ 726.905126] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 726.905126] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] updated_port = self._update_port( [ 726.905126] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 726.905126] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] _ensure_no_port_binding_failure(port) [ 726.905126] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 726.905126] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] raise exception.PortBindingFailed(port_id=port['id']) [ 726.905126] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] nova.exception.PortBindingFailed: Binding failed for port 403c6f43-6162-47c8-a025-754113fc33d6, please check neutron logs for more information. [ 726.905126] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] [ 726.905126] env[63088]: INFO nova.compute.manager [None req-03263d3e-caae-4f24-95aa-ac7aedd64259 tempest-ImagesOneServerTestJSON-44065314 tempest-ImagesOneServerTestJSON-44065314-project-member] [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] Terminating instance [ 726.906169] env[63088]: DEBUG oslo_concurrency.lockutils [None req-03263d3e-caae-4f24-95aa-ac7aedd64259 tempest-ImagesOneServerTestJSON-44065314 tempest-ImagesOneServerTestJSON-44065314-project-member] Acquiring lock "refresh_cache-c63c1740-fcec-47fa-a7b9-986b83c4abd9" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 726.946835] env[63088]: DEBUG nova.network.neutron [req-877a9f65-60cb-4d69-ae80-48c00de6bf2b req-96c4e89b-ce13-42ab-a63d-fe1f1ac5b46f service nova] [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 727.020776] env[63088]: DEBUG nova.network.neutron [req-877a9f65-60cb-4d69-ae80-48c00de6bf2b req-96c4e89b-ce13-42ab-a63d-fe1f1ac5b46f service nova] [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 727.209378] env[63088]: DEBUG nova.network.neutron [None req-0b29042d-57c3-47a9-a926-f67d685a4740 tempest-TenantUsagesTestJSON-1054626739 tempest-TenantUsagesTestJSON-1054626739-project-member] [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 727.267838] env[63088]: DEBUG nova.network.neutron [None req-0b29042d-57c3-47a9-a926-f67d685a4740 tempest-TenantUsagesTestJSON-1054626739 tempest-TenantUsagesTestJSON-1054626739-project-member] [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 727.411975] env[63088]: DEBUG nova.network.neutron [-] [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 727.523546] env[63088]: DEBUG oslo_concurrency.lockutils [req-877a9f65-60cb-4d69-ae80-48c00de6bf2b req-96c4e89b-ce13-42ab-a63d-fe1f1ac5b46f service nova] Releasing lock "refresh_cache-c63c1740-fcec-47fa-a7b9-986b83c4abd9" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 727.523987] env[63088]: DEBUG oslo_concurrency.lockutils [None req-03263d3e-caae-4f24-95aa-ac7aedd64259 tempest-ImagesOneServerTestJSON-44065314 tempest-ImagesOneServerTestJSON-44065314-project-member] Acquired lock "refresh_cache-c63c1740-fcec-47fa-a7b9-986b83c4abd9" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 727.524206] env[63088]: DEBUG nova.network.neutron [None req-03263d3e-caae-4f24-95aa-ac7aedd64259 tempest-ImagesOneServerTestJSON-44065314 tempest-ImagesOneServerTestJSON-44065314-project-member] [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 727.771163] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0b29042d-57c3-47a9-a926-f67d685a4740 tempest-TenantUsagesTestJSON-1054626739 tempest-TenantUsagesTestJSON-1054626739-project-member] Releasing lock "refresh_cache-02da88f0-7e3a-4eea-a88d-59a990c69f82" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 727.771163] env[63088]: DEBUG nova.compute.manager [None req-0b29042d-57c3-47a9-a926-f67d685a4740 tempest-TenantUsagesTestJSON-1054626739 tempest-TenantUsagesTestJSON-1054626739-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 727.771268] env[63088]: DEBUG nova.compute.manager [None req-0b29042d-57c3-47a9-a926-f67d685a4740 tempest-TenantUsagesTestJSON-1054626739 tempest-TenantUsagesTestJSON-1054626739-project-member] [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 727.771608] env[63088]: DEBUG nova.network.neutron [None req-0b29042d-57c3-47a9-a926-f67d685a4740 tempest-TenantUsagesTestJSON-1054626739 tempest-TenantUsagesTestJSON-1054626739-project-member] [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 727.788656] env[63088]: DEBUG nova.network.neutron [None req-0b29042d-57c3-47a9-a926-f67d685a4740 tempest-TenantUsagesTestJSON-1054626739 tempest-TenantUsagesTestJSON-1054626739-project-member] [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 727.914955] env[63088]: INFO nova.compute.manager [-] [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] Took 1.19 seconds to deallocate network for instance. [ 727.917188] env[63088]: DEBUG nova.compute.claims [None req-7af44f64-0067-49a1-ac0c-3ee5b8047867 tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] Aborting claim: {{(pid=63088) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 727.917302] env[63088]: DEBUG oslo_concurrency.lockutils [None req-7af44f64-0067-49a1-ac0c-3ee5b8047867 tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 727.961939] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe93f15e-4413-4158-a7a0-9c1df8d282a4 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.969611] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2d73e3f-c5d8-4f66-93a4-1badc3d9f072 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.998518] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d186f45-f328-4ab5-973a-3ec9c457dd69 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.005525] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fde1a53-5db7-46df-a24a-5ce9519be04e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.018135] env[63088]: DEBUG nova.compute.provider_tree [None req-a6077124-74e2-4777-9beb-9998c6275315 tempest-ServersTestBootFromVolume-2105176458 tempest-ServersTestBootFromVolume-2105176458-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 728.041567] env[63088]: DEBUG nova.network.neutron [None req-03263d3e-caae-4f24-95aa-ac7aedd64259 tempest-ImagesOneServerTestJSON-44065314 tempest-ImagesOneServerTestJSON-44065314-project-member] [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 728.119171] env[63088]: DEBUG nova.network.neutron [None req-03263d3e-caae-4f24-95aa-ac7aedd64259 tempest-ImagesOneServerTestJSON-44065314 tempest-ImagesOneServerTestJSON-44065314-project-member] [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 728.291571] env[63088]: DEBUG nova.network.neutron [None req-0b29042d-57c3-47a9-a926-f67d685a4740 tempest-TenantUsagesTestJSON-1054626739 tempest-TenantUsagesTestJSON-1054626739-project-member] [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 728.521583] env[63088]: DEBUG nova.scheduler.client.report [None req-a6077124-74e2-4777-9beb-9998c6275315 tempest-ServersTestBootFromVolume-2105176458 tempest-ServersTestBootFromVolume-2105176458-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 728.545047] env[63088]: DEBUG nova.compute.manager [req-ac1dcf34-e781-4479-af78-962d92f444f9 req-a51c31e6-886b-4bda-accd-01f7a45e644e service nova] [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] Received event network-vif-deleted-403c6f43-6162-47c8-a025-754113fc33d6 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 728.622263] env[63088]: DEBUG oslo_concurrency.lockutils [None req-03263d3e-caae-4f24-95aa-ac7aedd64259 tempest-ImagesOneServerTestJSON-44065314 tempest-ImagesOneServerTestJSON-44065314-project-member] Releasing lock "refresh_cache-c63c1740-fcec-47fa-a7b9-986b83c4abd9" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 728.622637] env[63088]: DEBUG nova.compute.manager [None req-03263d3e-caae-4f24-95aa-ac7aedd64259 tempest-ImagesOneServerTestJSON-44065314 tempest-ImagesOneServerTestJSON-44065314-project-member] [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 728.622834] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-03263d3e-caae-4f24-95aa-ac7aedd64259 tempest-ImagesOneServerTestJSON-44065314 tempest-ImagesOneServerTestJSON-44065314-project-member] [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 728.623150] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7ca2f685-87f7-424a-844e-8409199b3d92 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.632377] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35aec2f9-bf13-4e24-a21b-74db703c10df {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.656043] env[63088]: WARNING nova.virt.vmwareapi.vmops [None req-03263d3e-caae-4f24-95aa-ac7aedd64259 tempest-ImagesOneServerTestJSON-44065314 tempest-ImagesOneServerTestJSON-44065314-project-member] [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c63c1740-fcec-47fa-a7b9-986b83c4abd9 could not be found. [ 728.656384] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-03263d3e-caae-4f24-95aa-ac7aedd64259 tempest-ImagesOneServerTestJSON-44065314 tempest-ImagesOneServerTestJSON-44065314-project-member] [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 728.656675] env[63088]: INFO nova.compute.manager [None req-03263d3e-caae-4f24-95aa-ac7aedd64259 tempest-ImagesOneServerTestJSON-44065314 tempest-ImagesOneServerTestJSON-44065314-project-member] [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] Took 0.03 seconds to destroy the instance on the hypervisor. [ 728.657073] env[63088]: DEBUG oslo.service.loopingcall [None req-03263d3e-caae-4f24-95aa-ac7aedd64259 tempest-ImagesOneServerTestJSON-44065314 tempest-ImagesOneServerTestJSON-44065314-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 728.657274] env[63088]: DEBUG nova.compute.manager [-] [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 728.657372] env[63088]: DEBUG nova.network.neutron [-] [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 728.671102] env[63088]: DEBUG nova.network.neutron [-] [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 728.794507] env[63088]: INFO nova.compute.manager [None req-0b29042d-57c3-47a9-a926-f67d685a4740 tempest-TenantUsagesTestJSON-1054626739 tempest-TenantUsagesTestJSON-1054626739-project-member] [instance: 02da88f0-7e3a-4eea-a88d-59a990c69f82] Took 1.02 seconds to deallocate network for instance. [ 729.031026] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a6077124-74e2-4777-9beb-9998c6275315 tempest-ServersTestBootFromVolume-2105176458 tempest-ServersTestBootFromVolume-2105176458-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.350s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 729.031026] env[63088]: DEBUG nova.compute.manager [None req-a6077124-74e2-4777-9beb-9998c6275315 tempest-ServersTestBootFromVolume-2105176458 tempest-ServersTestBootFromVolume-2105176458-project-member] [instance: 407db84f-3322-42bc-baac-5528e856af8d] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 729.033274] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3b417ae5-c397-4c95-9150-090b3a6260b8 tempest-FloatingIPsAssociationTestJSON-911050413 tempest-FloatingIPsAssociationTestJSON-911050413-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.402s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 729.034591] env[63088]: INFO nova.compute.claims [None req-3b417ae5-c397-4c95-9150-090b3a6260b8 tempest-FloatingIPsAssociationTestJSON-911050413 tempest-FloatingIPsAssociationTestJSON-911050413-project-member] [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 729.172997] env[63088]: DEBUG nova.network.neutron [-] [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 729.539487] env[63088]: DEBUG nova.compute.utils [None req-a6077124-74e2-4777-9beb-9998c6275315 tempest-ServersTestBootFromVolume-2105176458 tempest-ServersTestBootFromVolume-2105176458-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 729.542808] env[63088]: DEBUG nova.compute.manager [None req-a6077124-74e2-4777-9beb-9998c6275315 tempest-ServersTestBootFromVolume-2105176458 tempest-ServersTestBootFromVolume-2105176458-project-member] [instance: 407db84f-3322-42bc-baac-5528e856af8d] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 729.542974] env[63088]: DEBUG nova.network.neutron [None req-a6077124-74e2-4777-9beb-9998c6275315 tempest-ServersTestBootFromVolume-2105176458 tempest-ServersTestBootFromVolume-2105176458-project-member] [instance: 407db84f-3322-42bc-baac-5528e856af8d] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 729.591518] env[63088]: DEBUG nova.policy [None req-a6077124-74e2-4777-9beb-9998c6275315 tempest-ServersTestBootFromVolume-2105176458 tempest-ServersTestBootFromVolume-2105176458-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7025184341864e1abc9a25b6ed57c1a4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '352fcbf280e74b8b83f8a5ad047a9dd3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 729.675129] env[63088]: INFO nova.compute.manager [-] [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] Took 1.02 seconds to deallocate network for instance. [ 729.677810] env[63088]: DEBUG nova.compute.claims [None req-03263d3e-caae-4f24-95aa-ac7aedd64259 tempest-ImagesOneServerTestJSON-44065314 tempest-ImagesOneServerTestJSON-44065314-project-member] [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] Aborting claim: {{(pid=63088) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 729.677983] env[63088]: DEBUG oslo_concurrency.lockutils [None req-03263d3e-caae-4f24-95aa-ac7aedd64259 tempest-ImagesOneServerTestJSON-44065314 tempest-ImagesOneServerTestJSON-44065314-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 729.830560] env[63088]: INFO nova.scheduler.client.report [None req-0b29042d-57c3-47a9-a926-f67d685a4740 tempest-TenantUsagesTestJSON-1054626739 tempest-TenantUsagesTestJSON-1054626739-project-member] Deleted allocations for instance 02da88f0-7e3a-4eea-a88d-59a990c69f82 [ 729.946109] env[63088]: DEBUG nova.network.neutron [None req-a6077124-74e2-4777-9beb-9998c6275315 tempest-ServersTestBootFromVolume-2105176458 tempest-ServersTestBootFromVolume-2105176458-project-member] [instance: 407db84f-3322-42bc-baac-5528e856af8d] Successfully created port: 8193ccbb-9d49-4c2e-a0b5-7a90dba1cb62 {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 730.043782] env[63088]: DEBUG nova.compute.manager [None req-a6077124-74e2-4777-9beb-9998c6275315 tempest-ServersTestBootFromVolume-2105176458 tempest-ServersTestBootFromVolume-2105176458-project-member] [instance: 407db84f-3322-42bc-baac-5528e856af8d] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 730.305052] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fab73f81-28bf-4cac-8691-9711f4a29e5e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.314878] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a42806bd-ccc5-4cdb-adc4-3cea16bf4f56 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.347392] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0b29042d-57c3-47a9-a926-f67d685a4740 tempest-TenantUsagesTestJSON-1054626739 tempest-TenantUsagesTestJSON-1054626739-project-member] Lock "02da88f0-7e3a-4eea-a88d-59a990c69f82" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 135.749s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 730.353018] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5360fc1-b7b2-4dff-ad8a-f75b4644a747 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.358077] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49feb57f-3ea9-4eb7-a693-cda6cd2cde41 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.373760] env[63088]: DEBUG nova.compute.provider_tree [None req-3b417ae5-c397-4c95-9150-090b3a6260b8 tempest-FloatingIPsAssociationTestJSON-911050413 tempest-FloatingIPsAssociationTestJSON-911050413-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 730.554060] env[63088]: INFO nova.virt.block_device [None req-a6077124-74e2-4777-9beb-9998c6275315 tempest-ServersTestBootFromVolume-2105176458 tempest-ServersTestBootFromVolume-2105176458-project-member] [instance: 407db84f-3322-42bc-baac-5528e856af8d] Booting with volume f7647177-d1ec-4638-8eb5-ab37e2a44684 at /dev/sda [ 730.604790] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e5386a9f-7a5d-4f56-987a-4c05b1a9e88d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.613518] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1245f65-e6e8-43f7-a163-60c48447e6ff {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.637265] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e6737a20-70d9-4975-8231-697a3f650ee5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.644953] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d398ad98-3f7d-4d9a-a2da-5776551e987e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.668064] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06fb6063-9dce-4a98-8151-54af84f05b2d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.671600] env[63088]: DEBUG nova.compute.manager [req-ea2dbc0b-6302-4242-80b0-037b552f27af req-8935229a-b6bc-46c3-9ae1-d2590323b978 service nova] [instance: 407db84f-3322-42bc-baac-5528e856af8d] Received event network-changed-8193ccbb-9d49-4c2e-a0b5-7a90dba1cb62 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 730.671760] env[63088]: DEBUG nova.compute.manager [req-ea2dbc0b-6302-4242-80b0-037b552f27af req-8935229a-b6bc-46c3-9ae1-d2590323b978 service nova] [instance: 407db84f-3322-42bc-baac-5528e856af8d] Refreshing instance network info cache due to event network-changed-8193ccbb-9d49-4c2e-a0b5-7a90dba1cb62. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 730.671926] env[63088]: DEBUG oslo_concurrency.lockutils [req-ea2dbc0b-6302-4242-80b0-037b552f27af req-8935229a-b6bc-46c3-9ae1-d2590323b978 service nova] Acquiring lock "refresh_cache-407db84f-3322-42bc-baac-5528e856af8d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 730.672047] env[63088]: DEBUG oslo_concurrency.lockutils [req-ea2dbc0b-6302-4242-80b0-037b552f27af req-8935229a-b6bc-46c3-9ae1-d2590323b978 service nova] Acquired lock "refresh_cache-407db84f-3322-42bc-baac-5528e856af8d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 730.672101] env[63088]: DEBUG nova.network.neutron [req-ea2dbc0b-6302-4242-80b0-037b552f27af req-8935229a-b6bc-46c3-9ae1-d2590323b978 service nova] [instance: 407db84f-3322-42bc-baac-5528e856af8d] Refreshing network info cache for port 8193ccbb-9d49-4c2e-a0b5-7a90dba1cb62 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 730.678913] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b6ed6ac-c43b-4638-ad77-dbfec25b8a72 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.693389] env[63088]: DEBUG nova.virt.block_device [None req-a6077124-74e2-4777-9beb-9998c6275315 tempest-ServersTestBootFromVolume-2105176458 tempest-ServersTestBootFromVolume-2105176458-project-member] [instance: 407db84f-3322-42bc-baac-5528e856af8d] Updating existing volume attachment record: 24b55eff-5e2b-4ee0-882d-f5adec0f26ae {{(pid=63088) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 730.857300] env[63088]: DEBUG nova.compute.manager [None req-d012c7de-73d3-432e-8cc9-9d2f0456777d tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 730.877669] env[63088]: DEBUG nova.scheduler.client.report [None req-3b417ae5-c397-4c95-9150-090b3a6260b8 tempest-FloatingIPsAssociationTestJSON-911050413 tempest-FloatingIPsAssociationTestJSON-911050413-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 730.891834] env[63088]: ERROR nova.compute.manager [None req-a6077124-74e2-4777-9beb-9998c6275315 tempest-ServersTestBootFromVolume-2105176458 tempest-ServersTestBootFromVolume-2105176458-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8193ccbb-9d49-4c2e-a0b5-7a90dba1cb62, please check neutron logs for more information. [ 730.891834] env[63088]: ERROR nova.compute.manager Traceback (most recent call last): [ 730.891834] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 730.891834] env[63088]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 730.891834] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 730.891834] env[63088]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 730.891834] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 730.891834] env[63088]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 730.891834] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 730.891834] env[63088]: ERROR nova.compute.manager self.force_reraise() [ 730.891834] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 730.891834] env[63088]: ERROR nova.compute.manager raise self.value [ 730.891834] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 730.891834] env[63088]: ERROR nova.compute.manager updated_port = self._update_port( [ 730.891834] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 730.891834] env[63088]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 730.892415] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 730.892415] env[63088]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 730.892415] env[63088]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8193ccbb-9d49-4c2e-a0b5-7a90dba1cb62, please check neutron logs for more information. [ 730.892415] env[63088]: ERROR nova.compute.manager [ 730.892415] env[63088]: Traceback (most recent call last): [ 730.892415] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 730.892415] env[63088]: listener.cb(fileno) [ 730.892415] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 730.892415] env[63088]: result = function(*args, **kwargs) [ 730.892415] env[63088]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 730.892415] env[63088]: return func(*args, **kwargs) [ 730.892415] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 730.892415] env[63088]: raise e [ 730.892415] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 730.892415] env[63088]: nwinfo = self.network_api.allocate_for_instance( [ 730.892415] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 730.892415] env[63088]: created_port_ids = self._update_ports_for_instance( [ 730.892415] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 730.892415] env[63088]: with excutils.save_and_reraise_exception(): [ 730.892415] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 730.892415] env[63088]: self.force_reraise() [ 730.892415] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 730.892415] env[63088]: raise self.value [ 730.892415] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 730.892415] env[63088]: updated_port = self._update_port( [ 730.892415] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 730.892415] env[63088]: _ensure_no_port_binding_failure(port) [ 730.892415] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 730.892415] env[63088]: raise exception.PortBindingFailed(port_id=port['id']) [ 730.893261] env[63088]: nova.exception.PortBindingFailed: Binding failed for port 8193ccbb-9d49-4c2e-a0b5-7a90dba1cb62, please check neutron logs for more information. [ 730.893261] env[63088]: Removing descriptor: 17 [ 731.190456] env[63088]: DEBUG nova.network.neutron [req-ea2dbc0b-6302-4242-80b0-037b552f27af req-8935229a-b6bc-46c3-9ae1-d2590323b978 service nova] [instance: 407db84f-3322-42bc-baac-5528e856af8d] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 731.277703] env[63088]: DEBUG nova.network.neutron [req-ea2dbc0b-6302-4242-80b0-037b552f27af req-8935229a-b6bc-46c3-9ae1-d2590323b978 service nova] [instance: 407db84f-3322-42bc-baac-5528e856af8d] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 731.377558] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d012c7de-73d3-432e-8cc9-9d2f0456777d tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 731.388419] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3b417ae5-c397-4c95-9150-090b3a6260b8 tempest-FloatingIPsAssociationTestJSON-911050413 tempest-FloatingIPsAssociationTestJSON-911050413-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.355s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 731.388967] env[63088]: DEBUG nova.compute.manager [None req-3b417ae5-c397-4c95-9150-090b3a6260b8 tempest-FloatingIPsAssociationTestJSON-911050413 tempest-FloatingIPsAssociationTestJSON-911050413-project-member] [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 731.392232] env[63088]: DEBUG oslo_concurrency.lockutils [None req-fcd5880e-5b8a-42fa-b638-13cd9836ac96 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.919s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 731.780657] env[63088]: DEBUG oslo_concurrency.lockutils [req-ea2dbc0b-6302-4242-80b0-037b552f27af req-8935229a-b6bc-46c3-9ae1-d2590323b978 service nova] Releasing lock "refresh_cache-407db84f-3322-42bc-baac-5528e856af8d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 731.893934] env[63088]: DEBUG nova.compute.utils [None req-3b417ae5-c397-4c95-9150-090b3a6260b8 tempest-FloatingIPsAssociationTestJSON-911050413 tempest-FloatingIPsAssociationTestJSON-911050413-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 731.895339] env[63088]: DEBUG nova.compute.manager [None req-3b417ae5-c397-4c95-9150-090b3a6260b8 tempest-FloatingIPsAssociationTestJSON-911050413 tempest-FloatingIPsAssociationTestJSON-911050413-project-member] [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 731.895516] env[63088]: DEBUG nova.network.neutron [None req-3b417ae5-c397-4c95-9150-090b3a6260b8 tempest-FloatingIPsAssociationTestJSON-911050413 tempest-FloatingIPsAssociationTestJSON-911050413-project-member] [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 731.952319] env[63088]: DEBUG nova.policy [None req-3b417ae5-c397-4c95-9150-090b3a6260b8 tempest-FloatingIPsAssociationTestJSON-911050413 tempest-FloatingIPsAssociationTestJSON-911050413-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4ebb2ea439fb496d952e0262de28f6cb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e0c8e3e47af14408bd4757157bf3ecd0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 732.179339] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6eea2d42-f563-4bbc-a817-6075fe10cff3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.187486] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95ca90b6-d972-40de-af45-c0cb25afe888 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.222789] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7678e47-b632-4d70-890c-d525a6926b29 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.231257] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b384cab5-24aa-4006-82f9-421d50c0d76d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.245997] env[63088]: DEBUG nova.compute.provider_tree [None req-fcd5880e-5b8a-42fa-b638-13cd9836ac96 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 732.248419] env[63088]: DEBUG nova.network.neutron [None req-3b417ae5-c397-4c95-9150-090b3a6260b8 tempest-FloatingIPsAssociationTestJSON-911050413 tempest-FloatingIPsAssociationTestJSON-911050413-project-member] [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] Successfully created port: eeaacdbe-07d9-4cc2-8775-873e396bf287 {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 732.400871] env[63088]: DEBUG nova.compute.manager [None req-3b417ae5-c397-4c95-9150-090b3a6260b8 tempest-FloatingIPsAssociationTestJSON-911050413 tempest-FloatingIPsAssociationTestJSON-911050413-project-member] [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 732.705828] env[63088]: DEBUG nova.compute.manager [req-4749af89-df78-428c-bfdb-9e650c05fa7f req-7f98db6d-4c28-436a-80da-e5d9856ed916 service nova] [instance: 407db84f-3322-42bc-baac-5528e856af8d] Received event network-vif-deleted-8193ccbb-9d49-4c2e-a0b5-7a90dba1cb62 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 732.751955] env[63088]: DEBUG nova.scheduler.client.report [None req-fcd5880e-5b8a-42fa-b638-13cd9836ac96 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 732.802741] env[63088]: DEBUG nova.compute.manager [None req-a6077124-74e2-4777-9beb-9998c6275315 tempest-ServersTestBootFromVolume-2105176458 tempest-ServersTestBootFromVolume-2105176458-project-member] [instance: 407db84f-3322-42bc-baac-5528e856af8d] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 732.803362] env[63088]: DEBUG nova.virt.hardware [None req-a6077124-74e2-4777-9beb-9998c6275315 tempest-ServersTestBootFromVolume-2105176458 tempest-ServersTestBootFromVolume-2105176458-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 732.803642] env[63088]: DEBUG nova.virt.hardware [None req-a6077124-74e2-4777-9beb-9998c6275315 tempest-ServersTestBootFromVolume-2105176458 tempest-ServersTestBootFromVolume-2105176458-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 732.803854] env[63088]: DEBUG nova.virt.hardware [None req-a6077124-74e2-4777-9beb-9998c6275315 tempest-ServersTestBootFromVolume-2105176458 tempest-ServersTestBootFromVolume-2105176458-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 732.804125] env[63088]: DEBUG nova.virt.hardware [None req-a6077124-74e2-4777-9beb-9998c6275315 tempest-ServersTestBootFromVolume-2105176458 tempest-ServersTestBootFromVolume-2105176458-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 732.804308] env[63088]: DEBUG nova.virt.hardware [None req-a6077124-74e2-4777-9beb-9998c6275315 tempest-ServersTestBootFromVolume-2105176458 tempest-ServersTestBootFromVolume-2105176458-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 732.804464] env[63088]: DEBUG nova.virt.hardware [None req-a6077124-74e2-4777-9beb-9998c6275315 tempest-ServersTestBootFromVolume-2105176458 tempest-ServersTestBootFromVolume-2105176458-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 732.804666] env[63088]: DEBUG nova.virt.hardware [None req-a6077124-74e2-4777-9beb-9998c6275315 tempest-ServersTestBootFromVolume-2105176458 tempest-ServersTestBootFromVolume-2105176458-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 732.804850] env[63088]: DEBUG nova.virt.hardware [None req-a6077124-74e2-4777-9beb-9998c6275315 tempest-ServersTestBootFromVolume-2105176458 tempest-ServersTestBootFromVolume-2105176458-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 732.805077] env[63088]: DEBUG nova.virt.hardware [None req-a6077124-74e2-4777-9beb-9998c6275315 tempest-ServersTestBootFromVolume-2105176458 tempest-ServersTestBootFromVolume-2105176458-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 732.805300] env[63088]: DEBUG nova.virt.hardware [None req-a6077124-74e2-4777-9beb-9998c6275315 tempest-ServersTestBootFromVolume-2105176458 tempest-ServersTestBootFromVolume-2105176458-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 732.805525] env[63088]: DEBUG nova.virt.hardware [None req-a6077124-74e2-4777-9beb-9998c6275315 tempest-ServersTestBootFromVolume-2105176458 tempest-ServersTestBootFromVolume-2105176458-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 732.806495] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-764e18ce-c768-4225-8f59-c21605b4e4da {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.816828] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eef7c431-112d-4f41-b28c-557423e830d9 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.830700] env[63088]: ERROR nova.compute.manager [None req-a6077124-74e2-4777-9beb-9998c6275315 tempest-ServersTestBootFromVolume-2105176458 tempest-ServersTestBootFromVolume-2105176458-project-member] [instance: 407db84f-3322-42bc-baac-5528e856af8d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8193ccbb-9d49-4c2e-a0b5-7a90dba1cb62, please check neutron logs for more information. [ 732.830700] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] Traceback (most recent call last): [ 732.830700] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 732.830700] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] yield resources [ 732.830700] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 732.830700] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] self.driver.spawn(context, instance, image_meta, [ 732.830700] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 732.830700] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 732.830700] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 732.830700] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] vm_ref = self.build_virtual_machine(instance, [ 732.830700] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 732.831092] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] vif_infos = vmwarevif.get_vif_info(self._session, [ 732.831092] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 732.831092] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] for vif in network_info: [ 732.831092] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 732.831092] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] return self._sync_wrapper(fn, *args, **kwargs) [ 732.831092] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 732.831092] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] self.wait() [ 732.831092] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 732.831092] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] self[:] = self._gt.wait() [ 732.831092] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 732.831092] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] return self._exit_event.wait() [ 732.831092] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 732.831092] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] current.throw(*self._exc) [ 732.831465] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 732.831465] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] result = function(*args, **kwargs) [ 732.831465] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 732.831465] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] return func(*args, **kwargs) [ 732.831465] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 732.831465] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] raise e [ 732.831465] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 732.831465] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] nwinfo = self.network_api.allocate_for_instance( [ 732.831465] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 732.831465] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] created_port_ids = self._update_ports_for_instance( [ 732.831465] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 732.831465] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] with excutils.save_and_reraise_exception(): [ 732.831465] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 732.831823] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] self.force_reraise() [ 732.831823] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 732.831823] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] raise self.value [ 732.831823] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 732.831823] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] updated_port = self._update_port( [ 732.831823] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 732.831823] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] _ensure_no_port_binding_failure(port) [ 732.831823] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 732.831823] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] raise exception.PortBindingFailed(port_id=port['id']) [ 732.831823] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] nova.exception.PortBindingFailed: Binding failed for port 8193ccbb-9d49-4c2e-a0b5-7a90dba1cb62, please check neutron logs for more information. [ 732.831823] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] [ 732.831823] env[63088]: INFO nova.compute.manager [None req-a6077124-74e2-4777-9beb-9998c6275315 tempest-ServersTestBootFromVolume-2105176458 tempest-ServersTestBootFromVolume-2105176458-project-member] [instance: 407db84f-3322-42bc-baac-5528e856af8d] Terminating instance [ 732.833103] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a6077124-74e2-4777-9beb-9998c6275315 tempest-ServersTestBootFromVolume-2105176458 tempest-ServersTestBootFromVolume-2105176458-project-member] Acquiring lock "refresh_cache-407db84f-3322-42bc-baac-5528e856af8d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 732.833335] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a6077124-74e2-4777-9beb-9998c6275315 tempest-ServersTestBootFromVolume-2105176458 tempest-ServersTestBootFromVolume-2105176458-project-member] Acquired lock "refresh_cache-407db84f-3322-42bc-baac-5528e856af8d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 732.833409] env[63088]: DEBUG nova.network.neutron [None req-a6077124-74e2-4777-9beb-9998c6275315 tempest-ServersTestBootFromVolume-2105176458 tempest-ServersTestBootFromVolume-2105176458-project-member] [instance: 407db84f-3322-42bc-baac-5528e856af8d] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 733.256798] env[63088]: ERROR nova.compute.manager [None req-3b417ae5-c397-4c95-9150-090b3a6260b8 tempest-FloatingIPsAssociationTestJSON-911050413 tempest-FloatingIPsAssociationTestJSON-911050413-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port eeaacdbe-07d9-4cc2-8775-873e396bf287, please check neutron logs for more information. [ 733.256798] env[63088]: ERROR nova.compute.manager Traceback (most recent call last): [ 733.256798] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 733.256798] env[63088]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 733.256798] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 733.256798] env[63088]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 733.256798] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 733.256798] env[63088]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 733.256798] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 733.256798] env[63088]: ERROR nova.compute.manager self.force_reraise() [ 733.256798] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 733.256798] env[63088]: ERROR nova.compute.manager raise self.value [ 733.256798] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 733.256798] env[63088]: ERROR nova.compute.manager updated_port = self._update_port( [ 733.256798] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 733.256798] env[63088]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 733.257408] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 733.257408] env[63088]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 733.257408] env[63088]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port eeaacdbe-07d9-4cc2-8775-873e396bf287, please check neutron logs for more information. [ 733.257408] env[63088]: ERROR nova.compute.manager [ 733.257408] env[63088]: Traceback (most recent call last): [ 733.257408] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 733.257408] env[63088]: listener.cb(fileno) [ 733.257408] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 733.257408] env[63088]: result = function(*args, **kwargs) [ 733.257408] env[63088]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 733.257408] env[63088]: return func(*args, **kwargs) [ 733.257408] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 733.257408] env[63088]: raise e [ 733.257408] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 733.257408] env[63088]: nwinfo = self.network_api.allocate_for_instance( [ 733.257408] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 733.257408] env[63088]: created_port_ids = self._update_ports_for_instance( [ 733.257408] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 733.257408] env[63088]: with excutils.save_and_reraise_exception(): [ 733.257408] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 733.257408] env[63088]: self.force_reraise() [ 733.257408] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 733.257408] env[63088]: raise self.value [ 733.257408] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 733.257408] env[63088]: updated_port = self._update_port( [ 733.257408] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 733.257408] env[63088]: _ensure_no_port_binding_failure(port) [ 733.257408] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 733.257408] env[63088]: raise exception.PortBindingFailed(port_id=port['id']) [ 733.258492] env[63088]: nova.exception.PortBindingFailed: Binding failed for port eeaacdbe-07d9-4cc2-8775-873e396bf287, please check neutron logs for more information. [ 733.258492] env[63088]: Removing descriptor: 17 [ 733.260967] env[63088]: DEBUG oslo_concurrency.lockutils [None req-fcd5880e-5b8a-42fa-b638-13cd9836ac96 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.869s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 733.261679] env[63088]: ERROR nova.compute.manager [None req-fcd5880e-5b8a-42fa-b638-13cd9836ac96 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 28873a04-8e37-4443-9cf4-585cce95276d, please check neutron logs for more information. [ 733.261679] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] Traceback (most recent call last): [ 733.261679] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 733.261679] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] self.driver.spawn(context, instance, image_meta, [ 733.261679] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 733.261679] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 733.261679] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 733.261679] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] vm_ref = self.build_virtual_machine(instance, [ 733.261679] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 733.261679] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] vif_infos = vmwarevif.get_vif_info(self._session, [ 733.261679] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 733.262033] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] for vif in network_info: [ 733.262033] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 733.262033] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] return self._sync_wrapper(fn, *args, **kwargs) [ 733.262033] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 733.262033] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] self.wait() [ 733.262033] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 733.262033] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] self[:] = self._gt.wait() [ 733.262033] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 733.262033] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] return self._exit_event.wait() [ 733.262033] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 733.262033] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] current.throw(*self._exc) [ 733.262033] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 733.262033] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] result = function(*args, **kwargs) [ 733.262436] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 733.262436] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] return func(*args, **kwargs) [ 733.262436] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 733.262436] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] raise e [ 733.262436] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 733.262436] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] nwinfo = self.network_api.allocate_for_instance( [ 733.262436] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 733.262436] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] created_port_ids = self._update_ports_for_instance( [ 733.262436] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 733.262436] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] with excutils.save_and_reraise_exception(): [ 733.262436] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 733.262436] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] self.force_reraise() [ 733.262436] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 733.262818] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] raise self.value [ 733.262818] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 733.262818] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] updated_port = self._update_port( [ 733.262818] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 733.262818] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] _ensure_no_port_binding_failure(port) [ 733.262818] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 733.262818] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] raise exception.PortBindingFailed(port_id=port['id']) [ 733.262818] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] nova.exception.PortBindingFailed: Binding failed for port 28873a04-8e37-4443-9cf4-585cce95276d, please check neutron logs for more information. [ 733.262818] env[63088]: ERROR nova.compute.manager [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] [ 733.262818] env[63088]: DEBUG nova.compute.utils [None req-fcd5880e-5b8a-42fa-b638-13cd9836ac96 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] Binding failed for port 28873a04-8e37-4443-9cf4-585cce95276d, please check neutron logs for more information. {{(pid=63088) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 733.263480] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6c63332a-ec25-45cc-9207-42e0be237d32 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.225s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 733.266702] env[63088]: DEBUG nova.compute.manager [None req-fcd5880e-5b8a-42fa-b638-13cd9836ac96 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] Build of instance c67b276f-07db-4cde-9f1c-efccc4d11c7d was re-scheduled: Binding failed for port 28873a04-8e37-4443-9cf4-585cce95276d, please check neutron logs for more information. {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 733.267145] env[63088]: DEBUG nova.compute.manager [None req-fcd5880e-5b8a-42fa-b638-13cd9836ac96 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] Unplugging VIFs for instance {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 733.267378] env[63088]: DEBUG oslo_concurrency.lockutils [None req-fcd5880e-5b8a-42fa-b638-13cd9836ac96 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Acquiring lock "refresh_cache-c67b276f-07db-4cde-9f1c-efccc4d11c7d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 733.267524] env[63088]: DEBUG oslo_concurrency.lockutils [None req-fcd5880e-5b8a-42fa-b638-13cd9836ac96 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Acquired lock "refresh_cache-c67b276f-07db-4cde-9f1c-efccc4d11c7d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 733.267681] env[63088]: DEBUG nova.network.neutron [None req-fcd5880e-5b8a-42fa-b638-13cd9836ac96 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 733.351177] env[63088]: DEBUG nova.network.neutron [None req-a6077124-74e2-4777-9beb-9998c6275315 tempest-ServersTestBootFromVolume-2105176458 tempest-ServersTestBootFromVolume-2105176458-project-member] [instance: 407db84f-3322-42bc-baac-5528e856af8d] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 733.408323] env[63088]: DEBUG nova.compute.manager [None req-3b417ae5-c397-4c95-9150-090b3a6260b8 tempest-FloatingIPsAssociationTestJSON-911050413 tempest-FloatingIPsAssociationTestJSON-911050413-project-member] [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 733.434030] env[63088]: DEBUG nova.virt.hardware [None req-3b417ae5-c397-4c95-9150-090b3a6260b8 tempest-FloatingIPsAssociationTestJSON-911050413 tempest-FloatingIPsAssociationTestJSON-911050413-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 733.434254] env[63088]: DEBUG nova.virt.hardware [None req-3b417ae5-c397-4c95-9150-090b3a6260b8 tempest-FloatingIPsAssociationTestJSON-911050413 tempest-FloatingIPsAssociationTestJSON-911050413-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 733.434419] env[63088]: DEBUG nova.virt.hardware [None req-3b417ae5-c397-4c95-9150-090b3a6260b8 tempest-FloatingIPsAssociationTestJSON-911050413 tempest-FloatingIPsAssociationTestJSON-911050413-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 733.434730] env[63088]: DEBUG nova.virt.hardware [None req-3b417ae5-c397-4c95-9150-090b3a6260b8 tempest-FloatingIPsAssociationTestJSON-911050413 tempest-FloatingIPsAssociationTestJSON-911050413-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 733.434884] env[63088]: DEBUG nova.virt.hardware [None req-3b417ae5-c397-4c95-9150-090b3a6260b8 tempest-FloatingIPsAssociationTestJSON-911050413 tempest-FloatingIPsAssociationTestJSON-911050413-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 733.435140] env[63088]: DEBUG nova.virt.hardware [None req-3b417ae5-c397-4c95-9150-090b3a6260b8 tempest-FloatingIPsAssociationTestJSON-911050413 tempest-FloatingIPsAssociationTestJSON-911050413-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 733.435404] env[63088]: DEBUG nova.virt.hardware [None req-3b417ae5-c397-4c95-9150-090b3a6260b8 tempest-FloatingIPsAssociationTestJSON-911050413 tempest-FloatingIPsAssociationTestJSON-911050413-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 733.435599] env[63088]: DEBUG nova.virt.hardware [None req-3b417ae5-c397-4c95-9150-090b3a6260b8 tempest-FloatingIPsAssociationTestJSON-911050413 tempest-FloatingIPsAssociationTestJSON-911050413-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 733.435778] env[63088]: DEBUG nova.virt.hardware [None req-3b417ae5-c397-4c95-9150-090b3a6260b8 tempest-FloatingIPsAssociationTestJSON-911050413 tempest-FloatingIPsAssociationTestJSON-911050413-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 733.435974] env[63088]: DEBUG nova.virt.hardware [None req-3b417ae5-c397-4c95-9150-090b3a6260b8 tempest-FloatingIPsAssociationTestJSON-911050413 tempest-FloatingIPsAssociationTestJSON-911050413-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 733.436207] env[63088]: DEBUG nova.virt.hardware [None req-3b417ae5-c397-4c95-9150-090b3a6260b8 tempest-FloatingIPsAssociationTestJSON-911050413 tempest-FloatingIPsAssociationTestJSON-911050413-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 733.437402] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae8aec5e-b417-45b7-967d-d0b6ae62d189 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.441025] env[63088]: DEBUG nova.network.neutron [None req-a6077124-74e2-4777-9beb-9998c6275315 tempest-ServersTestBootFromVolume-2105176458 tempest-ServersTestBootFromVolume-2105176458-project-member] [instance: 407db84f-3322-42bc-baac-5528e856af8d] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 733.447426] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5c17774-6872-40ea-bd2d-2488f61fbc93 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.461608] env[63088]: ERROR nova.compute.manager [None req-3b417ae5-c397-4c95-9150-090b3a6260b8 tempest-FloatingIPsAssociationTestJSON-911050413 tempest-FloatingIPsAssociationTestJSON-911050413-project-member] [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port eeaacdbe-07d9-4cc2-8775-873e396bf287, please check neutron logs for more information. [ 733.461608] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] Traceback (most recent call last): [ 733.461608] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 733.461608] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] yield resources [ 733.461608] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 733.461608] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] self.driver.spawn(context, instance, image_meta, [ 733.461608] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 733.461608] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] self._vmops.spawn(context, instance, image_meta, injected_files, [ 733.461608] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 733.461608] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] vm_ref = self.build_virtual_machine(instance, [ 733.461608] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 733.461925] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] vif_infos = vmwarevif.get_vif_info(self._session, [ 733.461925] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 733.461925] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] for vif in network_info: [ 733.461925] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 733.461925] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] return self._sync_wrapper(fn, *args, **kwargs) [ 733.461925] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 733.461925] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] self.wait() [ 733.461925] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 733.461925] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] self[:] = self._gt.wait() [ 733.461925] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 733.461925] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] return self._exit_event.wait() [ 733.461925] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 733.461925] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] current.throw(*self._exc) [ 733.463804] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 733.463804] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] result = function(*args, **kwargs) [ 733.463804] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 733.463804] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] return func(*args, **kwargs) [ 733.463804] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 733.463804] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] raise e [ 733.463804] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 733.463804] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] nwinfo = self.network_api.allocate_for_instance( [ 733.463804] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 733.463804] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] created_port_ids = self._update_ports_for_instance( [ 733.463804] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 733.463804] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] with excutils.save_and_reraise_exception(): [ 733.463804] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 733.464170] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] self.force_reraise() [ 733.464170] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 733.464170] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] raise self.value [ 733.464170] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 733.464170] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] updated_port = self._update_port( [ 733.464170] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 733.464170] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] _ensure_no_port_binding_failure(port) [ 733.464170] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 733.464170] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] raise exception.PortBindingFailed(port_id=port['id']) [ 733.464170] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] nova.exception.PortBindingFailed: Binding failed for port eeaacdbe-07d9-4cc2-8775-873e396bf287, please check neutron logs for more information. [ 733.464170] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] [ 733.464170] env[63088]: INFO nova.compute.manager [None req-3b417ae5-c397-4c95-9150-090b3a6260b8 tempest-FloatingIPsAssociationTestJSON-911050413 tempest-FloatingIPsAssociationTestJSON-911050413-project-member] [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] Terminating instance [ 733.464537] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3b417ae5-c397-4c95-9150-090b3a6260b8 tempest-FloatingIPsAssociationTestJSON-911050413 tempest-FloatingIPsAssociationTestJSON-911050413-project-member] Acquiring lock "refresh_cache-1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 733.464537] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3b417ae5-c397-4c95-9150-090b3a6260b8 tempest-FloatingIPsAssociationTestJSON-911050413 tempest-FloatingIPsAssociationTestJSON-911050413-project-member] Acquired lock "refresh_cache-1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 733.464692] env[63088]: DEBUG nova.network.neutron [None req-3b417ae5-c397-4c95-9150-090b3a6260b8 tempest-FloatingIPsAssociationTestJSON-911050413 tempest-FloatingIPsAssociationTestJSON-911050413-project-member] [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 733.786201] env[63088]: DEBUG nova.network.neutron [None req-fcd5880e-5b8a-42fa-b638-13cd9836ac96 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 733.865797] env[63088]: DEBUG nova.network.neutron [None req-fcd5880e-5b8a-42fa-b638-13cd9836ac96 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 733.947208] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a6077124-74e2-4777-9beb-9998c6275315 tempest-ServersTestBootFromVolume-2105176458 tempest-ServersTestBootFromVolume-2105176458-project-member] Releasing lock "refresh_cache-407db84f-3322-42bc-baac-5528e856af8d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 733.947208] env[63088]: DEBUG nova.compute.manager [None req-a6077124-74e2-4777-9beb-9998c6275315 tempest-ServersTestBootFromVolume-2105176458 tempest-ServersTestBootFromVolume-2105176458-project-member] [instance: 407db84f-3322-42bc-baac-5528e856af8d] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 733.947298] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8735d3fc-cea2-4c81-ba16-c74c0731729a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.955956] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0677e9be-91a0-4f84-a667-56f471f55ada {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.986775] env[63088]: WARNING nova.virt.vmwareapi.driver [None req-a6077124-74e2-4777-9beb-9998c6275315 tempest-ServersTestBootFromVolume-2105176458 tempest-ServersTestBootFromVolume-2105176458-project-member] [instance: 407db84f-3322-42bc-baac-5528e856af8d] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance 407db84f-3322-42bc-baac-5528e856af8d could not be found. [ 733.987017] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-a6077124-74e2-4777-9beb-9998c6275315 tempest-ServersTestBootFromVolume-2105176458 tempest-ServersTestBootFromVolume-2105176458-project-member] [instance: 407db84f-3322-42bc-baac-5528e856af8d] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 733.991398] env[63088]: DEBUG nova.network.neutron [None req-3b417ae5-c397-4c95-9150-090b3a6260b8 tempest-FloatingIPsAssociationTestJSON-911050413 tempest-FloatingIPsAssociationTestJSON-911050413-project-member] [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 733.993577] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-605996e2-fd31-4bff-a0f1-fc9a939c0694 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.001839] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-804bb4ba-2dbb-4337-88c6-bc7b54fffdd1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.024857] env[63088]: WARNING nova.virt.vmwareapi.vmops [None req-a6077124-74e2-4777-9beb-9998c6275315 tempest-ServersTestBootFromVolume-2105176458 tempest-ServersTestBootFromVolume-2105176458-project-member] [instance: 407db84f-3322-42bc-baac-5528e856af8d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 407db84f-3322-42bc-baac-5528e856af8d could not be found. [ 734.025091] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-a6077124-74e2-4777-9beb-9998c6275315 tempest-ServersTestBootFromVolume-2105176458 tempest-ServersTestBootFromVolume-2105176458-project-member] [instance: 407db84f-3322-42bc-baac-5528e856af8d] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 734.025731] env[63088]: INFO nova.compute.manager [None req-a6077124-74e2-4777-9beb-9998c6275315 tempest-ServersTestBootFromVolume-2105176458 tempest-ServersTestBootFromVolume-2105176458-project-member] [instance: 407db84f-3322-42bc-baac-5528e856af8d] Took 0.08 seconds to destroy the instance on the hypervisor. [ 734.025731] env[63088]: DEBUG oslo.service.loopingcall [None req-a6077124-74e2-4777-9beb-9998c6275315 tempest-ServersTestBootFromVolume-2105176458 tempest-ServersTestBootFromVolume-2105176458-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 734.025731] env[63088]: DEBUG nova.compute.manager [-] [instance: 407db84f-3322-42bc-baac-5528e856af8d] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 734.026189] env[63088]: DEBUG nova.network.neutron [-] [instance: 407db84f-3322-42bc-baac-5528e856af8d] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 734.045704] env[63088]: DEBUG nova.network.neutron [-] [instance: 407db84f-3322-42bc-baac-5528e856af8d] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 734.100570] env[63088]: DEBUG nova.network.neutron [None req-3b417ae5-c397-4c95-9150-090b3a6260b8 tempest-FloatingIPsAssociationTestJSON-911050413 tempest-FloatingIPsAssociationTestJSON-911050413-project-member] [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 734.107884] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6eaf46f9-7db2-414e-8e24-ab2dfb12b123 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.116519] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36dddccb-2dd0-49b1-95df-c7724ac4929a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.146637] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-823a24dc-d2a1-40c6-997b-4d1571f71d72 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.154275] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e65cc51a-695e-4868-bd31-3e7a5faca9e8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.167983] env[63088]: DEBUG nova.compute.provider_tree [None req-6c63332a-ec25-45cc-9207-42e0be237d32 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 734.368305] env[63088]: DEBUG oslo_concurrency.lockutils [None req-fcd5880e-5b8a-42fa-b638-13cd9836ac96 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Releasing lock "refresh_cache-c67b276f-07db-4cde-9f1c-efccc4d11c7d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 734.368547] env[63088]: DEBUG nova.compute.manager [None req-fcd5880e-5b8a-42fa-b638-13cd9836ac96 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 734.368815] env[63088]: DEBUG nova.compute.manager [None req-fcd5880e-5b8a-42fa-b638-13cd9836ac96 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 734.369044] env[63088]: DEBUG nova.network.neutron [None req-fcd5880e-5b8a-42fa-b638-13cd9836ac96 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 734.384094] env[63088]: DEBUG nova.network.neutron [None req-fcd5880e-5b8a-42fa-b638-13cd9836ac96 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 734.548502] env[63088]: DEBUG nova.network.neutron [-] [instance: 407db84f-3322-42bc-baac-5528e856af8d] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 734.602175] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3b417ae5-c397-4c95-9150-090b3a6260b8 tempest-FloatingIPsAssociationTestJSON-911050413 tempest-FloatingIPsAssociationTestJSON-911050413-project-member] Releasing lock "refresh_cache-1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 734.602578] env[63088]: DEBUG nova.compute.manager [None req-3b417ae5-c397-4c95-9150-090b3a6260b8 tempest-FloatingIPsAssociationTestJSON-911050413 tempest-FloatingIPsAssociationTestJSON-911050413-project-member] [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 734.602765] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-3b417ae5-c397-4c95-9150-090b3a6260b8 tempest-FloatingIPsAssociationTestJSON-911050413 tempest-FloatingIPsAssociationTestJSON-911050413-project-member] [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 734.603054] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-59741bd0-386a-4780-845b-6b4d8613667c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.611447] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df415406-5640-4f36-8922-7560479a511b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.632613] env[63088]: WARNING nova.virt.vmwareapi.vmops [None req-3b417ae5-c397-4c95-9150-090b3a6260b8 tempest-FloatingIPsAssociationTestJSON-911050413 tempest-FloatingIPsAssociationTestJSON-911050413-project-member] [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed could not be found. [ 734.632819] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-3b417ae5-c397-4c95-9150-090b3a6260b8 tempest-FloatingIPsAssociationTestJSON-911050413 tempest-FloatingIPsAssociationTestJSON-911050413-project-member] [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 734.632994] env[63088]: INFO nova.compute.manager [None req-3b417ae5-c397-4c95-9150-090b3a6260b8 tempest-FloatingIPsAssociationTestJSON-911050413 tempest-FloatingIPsAssociationTestJSON-911050413-project-member] [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] Took 0.03 seconds to destroy the instance on the hypervisor. [ 734.633245] env[63088]: DEBUG oslo.service.loopingcall [None req-3b417ae5-c397-4c95-9150-090b3a6260b8 tempest-FloatingIPsAssociationTestJSON-911050413 tempest-FloatingIPsAssociationTestJSON-911050413-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 734.633498] env[63088]: DEBUG nova.compute.manager [-] [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 734.633594] env[63088]: DEBUG nova.network.neutron [-] [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 734.646739] env[63088]: DEBUG nova.network.neutron [-] [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 734.672269] env[63088]: DEBUG nova.scheduler.client.report [None req-6c63332a-ec25-45cc-9207-42e0be237d32 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 734.732748] env[63088]: DEBUG nova.compute.manager [req-0cf74e5b-22ae-4f7a-a42a-7b622a21ec99 req-f28c8b1a-cdb3-4111-8fd3-053ae2478070 service nova] [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] Received event network-changed-eeaacdbe-07d9-4cc2-8775-873e396bf287 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 734.732929] env[63088]: DEBUG nova.compute.manager [req-0cf74e5b-22ae-4f7a-a42a-7b622a21ec99 req-f28c8b1a-cdb3-4111-8fd3-053ae2478070 service nova] [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] Refreshing instance network info cache due to event network-changed-eeaacdbe-07d9-4cc2-8775-873e396bf287. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 734.733171] env[63088]: DEBUG oslo_concurrency.lockutils [req-0cf74e5b-22ae-4f7a-a42a-7b622a21ec99 req-f28c8b1a-cdb3-4111-8fd3-053ae2478070 service nova] Acquiring lock "refresh_cache-1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 734.733272] env[63088]: DEBUG oslo_concurrency.lockutils [req-0cf74e5b-22ae-4f7a-a42a-7b622a21ec99 req-f28c8b1a-cdb3-4111-8fd3-053ae2478070 service nova] Acquired lock "refresh_cache-1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 734.733431] env[63088]: DEBUG nova.network.neutron [req-0cf74e5b-22ae-4f7a-a42a-7b622a21ec99 req-f28c8b1a-cdb3-4111-8fd3-053ae2478070 service nova] [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] Refreshing network info cache for port eeaacdbe-07d9-4cc2-8775-873e396bf287 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 734.886711] env[63088]: DEBUG nova.network.neutron [None req-fcd5880e-5b8a-42fa-b638-13cd9836ac96 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 735.051372] env[63088]: INFO nova.compute.manager [-] [instance: 407db84f-3322-42bc-baac-5528e856af8d] Took 1.03 seconds to deallocate network for instance. [ 735.149882] env[63088]: DEBUG nova.network.neutron [-] [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 735.176967] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6c63332a-ec25-45cc-9207-42e0be237d32 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.913s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 735.177704] env[63088]: ERROR nova.compute.manager [None req-6c63332a-ec25-45cc-9207-42e0be237d32 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c47ac4c5-a4a8-4766-bf26-05d24b003141, please check neutron logs for more information. [ 735.177704] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] Traceback (most recent call last): [ 735.177704] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 735.177704] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] self.driver.spawn(context, instance, image_meta, [ 735.177704] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 735.177704] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 735.177704] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 735.177704] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] vm_ref = self.build_virtual_machine(instance, [ 735.177704] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 735.177704] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] vif_infos = vmwarevif.get_vif_info(self._session, [ 735.177704] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 735.178112] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] for vif in network_info: [ 735.178112] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 735.178112] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] return self._sync_wrapper(fn, *args, **kwargs) [ 735.178112] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 735.178112] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] self.wait() [ 735.178112] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 735.178112] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] self[:] = self._gt.wait() [ 735.178112] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 735.178112] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] return self._exit_event.wait() [ 735.178112] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 735.178112] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] current.throw(*self._exc) [ 735.178112] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 735.178112] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] result = function(*args, **kwargs) [ 735.178502] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 735.178502] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] return func(*args, **kwargs) [ 735.178502] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 735.178502] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] raise e [ 735.178502] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 735.178502] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] nwinfo = self.network_api.allocate_for_instance( [ 735.178502] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 735.178502] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] created_port_ids = self._update_ports_for_instance( [ 735.178502] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 735.178502] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] with excutils.save_and_reraise_exception(): [ 735.178502] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 735.178502] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] self.force_reraise() [ 735.178502] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 735.178892] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] raise self.value [ 735.178892] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 735.178892] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] updated_port = self._update_port( [ 735.178892] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 735.178892] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] _ensure_no_port_binding_failure(port) [ 735.178892] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 735.178892] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] raise exception.PortBindingFailed(port_id=port['id']) [ 735.178892] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] nova.exception.PortBindingFailed: Binding failed for port c47ac4c5-a4a8-4766-bf26-05d24b003141, please check neutron logs for more information. [ 735.178892] env[63088]: ERROR nova.compute.manager [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] [ 735.178892] env[63088]: DEBUG nova.compute.utils [None req-6c63332a-ec25-45cc-9207-42e0be237d32 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] Binding failed for port c47ac4c5-a4a8-4766-bf26-05d24b003141, please check neutron logs for more information. {{(pid=63088) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 735.179542] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0c6bc062-cb90-4d50-aabf-8c0503771f7e tempest-InstanceActionsNegativeTestJSON-1719444358 tempest-InstanceActionsNegativeTestJSON-1719444358-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.792s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 735.186020] env[63088]: DEBUG nova.compute.manager [None req-6c63332a-ec25-45cc-9207-42e0be237d32 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] Build of instance 1219fe86-f218-4a2b-84ba-4a377829386d was re-scheduled: Binding failed for port c47ac4c5-a4a8-4766-bf26-05d24b003141, please check neutron logs for more information. {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 735.186020] env[63088]: DEBUG nova.compute.manager [None req-6c63332a-ec25-45cc-9207-42e0be237d32 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] Unplugging VIFs for instance {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 735.186020] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6c63332a-ec25-45cc-9207-42e0be237d32 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquiring lock "refresh_cache-1219fe86-f218-4a2b-84ba-4a377829386d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 735.186020] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6c63332a-ec25-45cc-9207-42e0be237d32 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquired lock "refresh_cache-1219fe86-f218-4a2b-84ba-4a377829386d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 735.186277] env[63088]: DEBUG nova.network.neutron [None req-6c63332a-ec25-45cc-9207-42e0be237d32 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 735.252423] env[63088]: DEBUG nova.network.neutron [req-0cf74e5b-22ae-4f7a-a42a-7b622a21ec99 req-f28c8b1a-cdb3-4111-8fd3-053ae2478070 service nova] [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 735.330875] env[63088]: DEBUG nova.network.neutron [req-0cf74e5b-22ae-4f7a-a42a-7b622a21ec99 req-f28c8b1a-cdb3-4111-8fd3-053ae2478070 service nova] [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 735.391050] env[63088]: INFO nova.compute.manager [None req-fcd5880e-5b8a-42fa-b638-13cd9836ac96 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] [instance: c67b276f-07db-4cde-9f1c-efccc4d11c7d] Took 1.02 seconds to deallocate network for instance. [ 735.609030] env[63088]: INFO nova.compute.manager [None req-a6077124-74e2-4777-9beb-9998c6275315 tempest-ServersTestBootFromVolume-2105176458 tempest-ServersTestBootFromVolume-2105176458-project-member] [instance: 407db84f-3322-42bc-baac-5528e856af8d] Took 0.56 seconds to detach 1 volumes for instance. [ 735.611579] env[63088]: DEBUG nova.compute.claims [None req-a6077124-74e2-4777-9beb-9998c6275315 tempest-ServersTestBootFromVolume-2105176458 tempest-ServersTestBootFromVolume-2105176458-project-member] [instance: 407db84f-3322-42bc-baac-5528e856af8d] Aborting claim: {{(pid=63088) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 735.611756] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a6077124-74e2-4777-9beb-9998c6275315 tempest-ServersTestBootFromVolume-2105176458 tempest-ServersTestBootFromVolume-2105176458-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 735.653130] env[63088]: INFO nova.compute.manager [-] [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] Took 1.02 seconds to deallocate network for instance. [ 735.654992] env[63088]: DEBUG nova.compute.claims [None req-3b417ae5-c397-4c95-9150-090b3a6260b8 tempest-FloatingIPsAssociationTestJSON-911050413 tempest-FloatingIPsAssociationTestJSON-911050413-project-member] [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] Aborting claim: {{(pid=63088) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 735.655174] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3b417ae5-c397-4c95-9150-090b3a6260b8 tempest-FloatingIPsAssociationTestJSON-911050413 tempest-FloatingIPsAssociationTestJSON-911050413-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 735.709569] env[63088]: DEBUG nova.network.neutron [None req-6c63332a-ec25-45cc-9207-42e0be237d32 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 735.807825] env[63088]: DEBUG nova.network.neutron [None req-6c63332a-ec25-45cc-9207-42e0be237d32 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 735.833332] env[63088]: DEBUG oslo_concurrency.lockutils [req-0cf74e5b-22ae-4f7a-a42a-7b622a21ec99 req-f28c8b1a-cdb3-4111-8fd3-053ae2478070 service nova] Releasing lock "refresh_cache-1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 735.833332] env[63088]: DEBUG nova.compute.manager [req-0cf74e5b-22ae-4f7a-a42a-7b622a21ec99 req-f28c8b1a-cdb3-4111-8fd3-053ae2478070 service nova] [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] Received event network-vif-deleted-eeaacdbe-07d9-4cc2-8775-873e396bf287 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 735.959475] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b558650a-8f8e-40c2-8c4b-1aa401066fe1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.967093] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7482414-9415-47d0-a253-5188d0a1657e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.998842] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3df50e72-49cf-46ed-bbbc-e43838da648b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.005904] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-858afdc8-aa45-497d-9ef2-564726e03faf {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.018701] env[63088]: DEBUG nova.compute.provider_tree [None req-0c6bc062-cb90-4d50-aabf-8c0503771f7e tempest-InstanceActionsNegativeTestJSON-1719444358 tempest-InstanceActionsNegativeTestJSON-1719444358-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 736.310722] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6c63332a-ec25-45cc-9207-42e0be237d32 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Releasing lock "refresh_cache-1219fe86-f218-4a2b-84ba-4a377829386d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 736.310975] env[63088]: DEBUG nova.compute.manager [None req-6c63332a-ec25-45cc-9207-42e0be237d32 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 736.311180] env[63088]: DEBUG nova.compute.manager [None req-6c63332a-ec25-45cc-9207-42e0be237d32 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 736.311351] env[63088]: DEBUG nova.network.neutron [None req-6c63332a-ec25-45cc-9207-42e0be237d32 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 736.326019] env[63088]: DEBUG nova.network.neutron [None req-6c63332a-ec25-45cc-9207-42e0be237d32 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 736.420434] env[63088]: INFO nova.scheduler.client.report [None req-fcd5880e-5b8a-42fa-b638-13cd9836ac96 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Deleted allocations for instance c67b276f-07db-4cde-9f1c-efccc4d11c7d [ 736.521603] env[63088]: DEBUG nova.scheduler.client.report [None req-0c6bc062-cb90-4d50-aabf-8c0503771f7e tempest-InstanceActionsNegativeTestJSON-1719444358 tempest-InstanceActionsNegativeTestJSON-1719444358-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 736.828308] env[63088]: DEBUG nova.network.neutron [None req-6c63332a-ec25-45cc-9207-42e0be237d32 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 736.927772] env[63088]: DEBUG oslo_concurrency.lockutils [None req-fcd5880e-5b8a-42fa-b638-13cd9836ac96 tempest-ListServerFiltersTestJSON-2131479764 tempest-ListServerFiltersTestJSON-2131479764-project-member] Lock "c67b276f-07db-4cde-9f1c-efccc4d11c7d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 141.233s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 737.026573] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0c6bc062-cb90-4d50-aabf-8c0503771f7e tempest-InstanceActionsNegativeTestJSON-1719444358 tempest-InstanceActionsNegativeTestJSON-1719444358-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.847s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 737.027221] env[63088]: ERROR nova.compute.manager [None req-0c6bc062-cb90-4d50-aabf-8c0503771f7e tempest-InstanceActionsNegativeTestJSON-1719444358 tempest-InstanceActionsNegativeTestJSON-1719444358-project-member] [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9c453a07-3eda-4075-81ce-a4516dc00de7, please check neutron logs for more information. [ 737.027221] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] Traceback (most recent call last): [ 737.027221] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 737.027221] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] self.driver.spawn(context, instance, image_meta, [ 737.027221] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 737.027221] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] self._vmops.spawn(context, instance, image_meta, injected_files, [ 737.027221] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 737.027221] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] vm_ref = self.build_virtual_machine(instance, [ 737.027221] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 737.027221] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] vif_infos = vmwarevif.get_vif_info(self._session, [ 737.027221] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 737.027568] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] for vif in network_info: [ 737.027568] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 737.027568] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] return self._sync_wrapper(fn, *args, **kwargs) [ 737.027568] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 737.027568] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] self.wait() [ 737.027568] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 737.027568] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] self[:] = self._gt.wait() [ 737.027568] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 737.027568] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] return self._exit_event.wait() [ 737.027568] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 737.027568] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] current.throw(*self._exc) [ 737.027568] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 737.027568] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] result = function(*args, **kwargs) [ 737.028090] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 737.028090] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] return func(*args, **kwargs) [ 737.028090] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 737.028090] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] raise e [ 737.028090] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 737.028090] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] nwinfo = self.network_api.allocate_for_instance( [ 737.028090] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 737.028090] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] created_port_ids = self._update_ports_for_instance( [ 737.028090] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 737.028090] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] with excutils.save_and_reraise_exception(): [ 737.028090] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 737.028090] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] self.force_reraise() [ 737.028090] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 737.028519] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] raise self.value [ 737.028519] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 737.028519] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] updated_port = self._update_port( [ 737.028519] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 737.028519] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] _ensure_no_port_binding_failure(port) [ 737.028519] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 737.028519] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] raise exception.PortBindingFailed(port_id=port['id']) [ 737.028519] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] nova.exception.PortBindingFailed: Binding failed for port 9c453a07-3eda-4075-81ce-a4516dc00de7, please check neutron logs for more information. [ 737.028519] env[63088]: ERROR nova.compute.manager [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] [ 737.028519] env[63088]: DEBUG nova.compute.utils [None req-0c6bc062-cb90-4d50-aabf-8c0503771f7e tempest-InstanceActionsNegativeTestJSON-1719444358 tempest-InstanceActionsNegativeTestJSON-1719444358-project-member] [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] Binding failed for port 9c453a07-3eda-4075-81ce-a4516dc00de7, please check neutron logs for more information. {{(pid=63088) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 737.029204] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a2a0ac90-0e7a-45d1-a070-f7ee6be01d68 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.200s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 737.032030] env[63088]: DEBUG nova.compute.manager [None req-0c6bc062-cb90-4d50-aabf-8c0503771f7e tempest-InstanceActionsNegativeTestJSON-1719444358 tempest-InstanceActionsNegativeTestJSON-1719444358-project-member] [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] Build of instance 20611ed4-97b3-4d6d-bef7-c13e1235c401 was re-scheduled: Binding failed for port 9c453a07-3eda-4075-81ce-a4516dc00de7, please check neutron logs for more information. {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 737.032446] env[63088]: DEBUG nova.compute.manager [None req-0c6bc062-cb90-4d50-aabf-8c0503771f7e tempest-InstanceActionsNegativeTestJSON-1719444358 tempest-InstanceActionsNegativeTestJSON-1719444358-project-member] [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] Unplugging VIFs for instance {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 737.032669] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0c6bc062-cb90-4d50-aabf-8c0503771f7e tempest-InstanceActionsNegativeTestJSON-1719444358 tempest-InstanceActionsNegativeTestJSON-1719444358-project-member] Acquiring lock "refresh_cache-20611ed4-97b3-4d6d-bef7-c13e1235c401" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 737.032811] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0c6bc062-cb90-4d50-aabf-8c0503771f7e tempest-InstanceActionsNegativeTestJSON-1719444358 tempest-InstanceActionsNegativeTestJSON-1719444358-project-member] Acquired lock "refresh_cache-20611ed4-97b3-4d6d-bef7-c13e1235c401" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 737.032967] env[63088]: DEBUG nova.network.neutron [None req-0c6bc062-cb90-4d50-aabf-8c0503771f7e tempest-InstanceActionsNegativeTestJSON-1719444358 tempest-InstanceActionsNegativeTestJSON-1719444358-project-member] [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 737.330818] env[63088]: INFO nova.compute.manager [None req-6c63332a-ec25-45cc-9207-42e0be237d32 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 1219fe86-f218-4a2b-84ba-4a377829386d] Took 1.02 seconds to deallocate network for instance. [ 737.430863] env[63088]: DEBUG nova.compute.manager [None req-bce00a2e-7fb0-44a9-a66e-261b7c3d5384 tempest-ServerActionsV293TestJSON-1088890325 tempest-ServerActionsV293TestJSON-1088890325-project-member] [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 737.554463] env[63088]: DEBUG nova.network.neutron [None req-0c6bc062-cb90-4d50-aabf-8c0503771f7e tempest-InstanceActionsNegativeTestJSON-1719444358 tempest-InstanceActionsNegativeTestJSON-1719444358-project-member] [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 737.691376] env[63088]: DEBUG nova.network.neutron [None req-0c6bc062-cb90-4d50-aabf-8c0503771f7e tempest-InstanceActionsNegativeTestJSON-1719444358 tempest-InstanceActionsNegativeTestJSON-1719444358-project-member] [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 737.803025] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72253196-bf02-4b40-8ba5-dcf4a48cf099 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.811580] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aec48fbd-db80-4bca-8403-bb332a6b5345 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.846039] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb5022f8-3c38-4905-8deb-a2d70a8fb338 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.855204] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e6d4a50-7eb3-4213-9ea5-364bc19763ba {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.869460] env[63088]: DEBUG nova.compute.provider_tree [None req-a2a0ac90-0e7a-45d1-a070-f7ee6be01d68 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 737.954377] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bce00a2e-7fb0-44a9-a66e-261b7c3d5384 tempest-ServerActionsV293TestJSON-1088890325 tempest-ServerActionsV293TestJSON-1088890325-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 738.194206] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0c6bc062-cb90-4d50-aabf-8c0503771f7e tempest-InstanceActionsNegativeTestJSON-1719444358 tempest-InstanceActionsNegativeTestJSON-1719444358-project-member] Releasing lock "refresh_cache-20611ed4-97b3-4d6d-bef7-c13e1235c401" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 738.194497] env[63088]: DEBUG nova.compute.manager [None req-0c6bc062-cb90-4d50-aabf-8c0503771f7e tempest-InstanceActionsNegativeTestJSON-1719444358 tempest-InstanceActionsNegativeTestJSON-1719444358-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 738.194497] env[63088]: DEBUG nova.compute.manager [None req-0c6bc062-cb90-4d50-aabf-8c0503771f7e tempest-InstanceActionsNegativeTestJSON-1719444358 tempest-InstanceActionsNegativeTestJSON-1719444358-project-member] [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 738.194752] env[63088]: DEBUG nova.network.neutron [None req-0c6bc062-cb90-4d50-aabf-8c0503771f7e tempest-InstanceActionsNegativeTestJSON-1719444358 tempest-InstanceActionsNegativeTestJSON-1719444358-project-member] [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 738.211705] env[63088]: DEBUG nova.network.neutron [None req-0c6bc062-cb90-4d50-aabf-8c0503771f7e tempest-InstanceActionsNegativeTestJSON-1719444358 tempest-InstanceActionsNegativeTestJSON-1719444358-project-member] [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 738.369437] env[63088]: INFO nova.scheduler.client.report [None req-6c63332a-ec25-45cc-9207-42e0be237d32 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Deleted allocations for instance 1219fe86-f218-4a2b-84ba-4a377829386d [ 738.376881] env[63088]: DEBUG nova.scheduler.client.report [None req-a2a0ac90-0e7a-45d1-a070-f7ee6be01d68 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 738.714414] env[63088]: DEBUG nova.network.neutron [None req-0c6bc062-cb90-4d50-aabf-8c0503771f7e tempest-InstanceActionsNegativeTestJSON-1719444358 tempest-InstanceActionsNegativeTestJSON-1719444358-project-member] [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 738.882047] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a2a0ac90-0e7a-45d1-a070-f7ee6be01d68 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.853s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 738.882709] env[63088]: ERROR nova.compute.manager [None req-a2a0ac90-0e7a-45d1-a070-f7ee6be01d68 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 69404326-63f0-4b81-8180-eae329f85f93, please check neutron logs for more information. [ 738.882709] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] Traceback (most recent call last): [ 738.882709] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 738.882709] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] self.driver.spawn(context, instance, image_meta, [ 738.882709] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 738.882709] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 738.882709] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 738.882709] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] vm_ref = self.build_virtual_machine(instance, [ 738.882709] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 738.882709] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] vif_infos = vmwarevif.get_vif_info(self._session, [ 738.882709] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 738.883022] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] for vif in network_info: [ 738.883022] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 738.883022] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] return self._sync_wrapper(fn, *args, **kwargs) [ 738.883022] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 738.883022] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] self.wait() [ 738.883022] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 738.883022] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] self[:] = self._gt.wait() [ 738.883022] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 738.883022] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] return self._exit_event.wait() [ 738.883022] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 738.883022] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] current.throw(*self._exc) [ 738.883022] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 738.883022] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] result = function(*args, **kwargs) [ 738.883343] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 738.883343] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] return func(*args, **kwargs) [ 738.883343] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 738.883343] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] raise e [ 738.883343] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 738.883343] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] nwinfo = self.network_api.allocate_for_instance( [ 738.883343] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 738.883343] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] created_port_ids = self._update_ports_for_instance( [ 738.883343] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 738.883343] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] with excutils.save_and_reraise_exception(): [ 738.883343] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 738.883343] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] self.force_reraise() [ 738.883343] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 738.883661] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] raise self.value [ 738.883661] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 738.883661] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] updated_port = self._update_port( [ 738.883661] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 738.883661] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] _ensure_no_port_binding_failure(port) [ 738.883661] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 738.883661] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] raise exception.PortBindingFailed(port_id=port['id']) [ 738.883661] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] nova.exception.PortBindingFailed: Binding failed for port 69404326-63f0-4b81-8180-eae329f85f93, please check neutron logs for more information. [ 738.883661] env[63088]: ERROR nova.compute.manager [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] [ 738.883661] env[63088]: DEBUG nova.compute.utils [None req-a2a0ac90-0e7a-45d1-a070-f7ee6be01d68 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] Binding failed for port 69404326-63f0-4b81-8180-eae329f85f93, please check neutron logs for more information. {{(pid=63088) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 738.885383] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6c63332a-ec25-45cc-9207-42e0be237d32 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Lock "1219fe86-f218-4a2b-84ba-4a377829386d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 140.870s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 738.885844] env[63088]: DEBUG nova.compute.manager [None req-a2a0ac90-0e7a-45d1-a070-f7ee6be01d68 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] Build of instance eb42c667-b0ae-4807-8f69-88dd671aa7d7 was re-scheduled: Binding failed for port 69404326-63f0-4b81-8180-eae329f85f93, please check neutron logs for more information. {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 738.886310] env[63088]: DEBUG nova.compute.manager [None req-a2a0ac90-0e7a-45d1-a070-f7ee6be01d68 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] Unplugging VIFs for instance {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 738.886580] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a2a0ac90-0e7a-45d1-a070-f7ee6be01d68 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquiring lock "refresh_cache-eb42c667-b0ae-4807-8f69-88dd671aa7d7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 738.886758] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a2a0ac90-0e7a-45d1-a070-f7ee6be01d68 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquired lock "refresh_cache-eb42c667-b0ae-4807-8f69-88dd671aa7d7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 738.886947] env[63088]: DEBUG nova.network.neutron [None req-a2a0ac90-0e7a-45d1-a070-f7ee6be01d68 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 738.888026] env[63088]: DEBUG oslo_concurrency.lockutils [None req-ae448e13-3c80-41d5-9b1f-e650107512d4 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.591s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 739.217741] env[63088]: INFO nova.compute.manager [None req-0c6bc062-cb90-4d50-aabf-8c0503771f7e tempest-InstanceActionsNegativeTestJSON-1719444358 tempest-InstanceActionsNegativeTestJSON-1719444358-project-member] [instance: 20611ed4-97b3-4d6d-bef7-c13e1235c401] Took 1.02 seconds to deallocate network for instance. [ 739.397747] env[63088]: DEBUG nova.compute.manager [None req-1874938e-086f-4c3c-9049-55ea0246719c tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 739.428706] env[63088]: DEBUG nova.network.neutron [None req-a2a0ac90-0e7a-45d1-a070-f7ee6be01d68 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 739.535171] env[63088]: DEBUG nova.network.neutron [None req-a2a0ac90-0e7a-45d1-a070-f7ee6be01d68 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 739.653404] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8d1c7e1-9499-481b-a4c9-fc87ef4e2cc1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.660715] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d917967e-4df5-4ad4-813f-c7995dbe9e21 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.689860] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d2e4d64-7e56-4fdd-a143-092176d19975 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.696853] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae9c1aa6-1f81-4a7c-ae1b-d64260d56f89 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.710784] env[63088]: DEBUG nova.compute.provider_tree [None req-ae448e13-3c80-41d5-9b1f-e650107512d4 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 739.918402] env[63088]: DEBUG oslo_concurrency.lockutils [None req-1874938e-086f-4c3c-9049-55ea0246719c tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 740.037939] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a2a0ac90-0e7a-45d1-a070-f7ee6be01d68 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Releasing lock "refresh_cache-eb42c667-b0ae-4807-8f69-88dd671aa7d7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 740.038172] env[63088]: DEBUG nova.compute.manager [None req-a2a0ac90-0e7a-45d1-a070-f7ee6be01d68 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 740.038360] env[63088]: DEBUG nova.compute.manager [None req-a2a0ac90-0e7a-45d1-a070-f7ee6be01d68 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 740.038528] env[63088]: DEBUG nova.network.neutron [None req-a2a0ac90-0e7a-45d1-a070-f7ee6be01d68 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 740.060841] env[63088]: DEBUG nova.network.neutron [None req-a2a0ac90-0e7a-45d1-a070-f7ee6be01d68 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 740.214460] env[63088]: DEBUG nova.scheduler.client.report [None req-ae448e13-3c80-41d5-9b1f-e650107512d4 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 740.247161] env[63088]: INFO nova.scheduler.client.report [None req-0c6bc062-cb90-4d50-aabf-8c0503771f7e tempest-InstanceActionsNegativeTestJSON-1719444358 tempest-InstanceActionsNegativeTestJSON-1719444358-project-member] Deleted allocations for instance 20611ed4-97b3-4d6d-bef7-c13e1235c401 [ 740.564051] env[63088]: DEBUG nova.network.neutron [None req-a2a0ac90-0e7a-45d1-a070-f7ee6be01d68 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 740.719461] env[63088]: DEBUG oslo_concurrency.lockutils [None req-ae448e13-3c80-41d5-9b1f-e650107512d4 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.831s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 740.720152] env[63088]: ERROR nova.compute.manager [None req-ae448e13-3c80-41d5-9b1f-e650107512d4 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c7c50755-fc74-476d-8236-768894ed1858, please check neutron logs for more information. [ 740.720152] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] Traceback (most recent call last): [ 740.720152] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 740.720152] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] self.driver.spawn(context, instance, image_meta, [ 740.720152] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 740.720152] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] self._vmops.spawn(context, instance, image_meta, injected_files, [ 740.720152] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 740.720152] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] vm_ref = self.build_virtual_machine(instance, [ 740.720152] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 740.720152] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] vif_infos = vmwarevif.get_vif_info(self._session, [ 740.720152] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 740.720590] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] for vif in network_info: [ 740.720590] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 740.720590] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] return self._sync_wrapper(fn, *args, **kwargs) [ 740.720590] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 740.720590] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] self.wait() [ 740.720590] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 740.720590] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] self[:] = self._gt.wait() [ 740.720590] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 740.720590] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] return self._exit_event.wait() [ 740.720590] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 740.720590] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] current.throw(*self._exc) [ 740.720590] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 740.720590] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] result = function(*args, **kwargs) [ 740.721108] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 740.721108] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] return func(*args, **kwargs) [ 740.721108] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 740.721108] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] raise e [ 740.721108] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 740.721108] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] nwinfo = self.network_api.allocate_for_instance( [ 740.721108] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 740.721108] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] created_port_ids = self._update_ports_for_instance( [ 740.721108] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 740.721108] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] with excutils.save_and_reraise_exception(): [ 740.721108] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 740.721108] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] self.force_reraise() [ 740.721108] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 740.721492] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] raise self.value [ 740.721492] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 740.721492] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] updated_port = self._update_port( [ 740.721492] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 740.721492] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] _ensure_no_port_binding_failure(port) [ 740.721492] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 740.721492] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] raise exception.PortBindingFailed(port_id=port['id']) [ 740.721492] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] nova.exception.PortBindingFailed: Binding failed for port c7c50755-fc74-476d-8236-768894ed1858, please check neutron logs for more information. [ 740.721492] env[63088]: ERROR nova.compute.manager [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] [ 740.721492] env[63088]: DEBUG nova.compute.utils [None req-ae448e13-3c80-41d5-9b1f-e650107512d4 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] Binding failed for port c7c50755-fc74-476d-8236-768894ed1858, please check neutron logs for more information. {{(pid=63088) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 740.722275] env[63088]: DEBUG oslo_concurrency.lockutils [None req-7af44f64-0067-49a1-ac0c-3ee5b8047867 tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.805s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 740.725338] env[63088]: DEBUG nova.compute.manager [None req-ae448e13-3c80-41d5-9b1f-e650107512d4 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] Build of instance 2b43620d-94ac-4c56-87df-4066f5d24088 was re-scheduled: Binding failed for port c7c50755-fc74-476d-8236-768894ed1858, please check neutron logs for more information. {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 740.725484] env[63088]: DEBUG nova.compute.manager [None req-ae448e13-3c80-41d5-9b1f-e650107512d4 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] Unplugging VIFs for instance {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 740.725715] env[63088]: DEBUG oslo_concurrency.lockutils [None req-ae448e13-3c80-41d5-9b1f-e650107512d4 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquiring lock "refresh_cache-2b43620d-94ac-4c56-87df-4066f5d24088" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 740.725862] env[63088]: DEBUG oslo_concurrency.lockutils [None req-ae448e13-3c80-41d5-9b1f-e650107512d4 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquired lock "refresh_cache-2b43620d-94ac-4c56-87df-4066f5d24088" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 740.726032] env[63088]: DEBUG nova.network.neutron [None req-ae448e13-3c80-41d5-9b1f-e650107512d4 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 740.756704] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0c6bc062-cb90-4d50-aabf-8c0503771f7e tempest-InstanceActionsNegativeTestJSON-1719444358 tempest-InstanceActionsNegativeTestJSON-1719444358-project-member] Lock "20611ed4-97b3-4d6d-bef7-c13e1235c401" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 142.130s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 741.067627] env[63088]: INFO nova.compute.manager [None req-a2a0ac90-0e7a-45d1-a070-f7ee6be01d68 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: eb42c667-b0ae-4807-8f69-88dd671aa7d7] Took 1.03 seconds to deallocate network for instance. [ 741.257398] env[63088]: DEBUG nova.network.neutron [None req-ae448e13-3c80-41d5-9b1f-e650107512d4 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 741.258529] env[63088]: DEBUG nova.compute.manager [None req-3edaa40f-e8ca-4f6f-a9f2-5744f490f65d tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 741.371272] env[63088]: DEBUG nova.network.neutron [None req-ae448e13-3c80-41d5-9b1f-e650107512d4 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 741.510057] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ecccc9c-b1cb-40c4-be4c-7a6e55a1fda3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.518459] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73f812ee-bf73-48af-ac20-a7e4c222d522 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.549118] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86f2868e-f529-46df-887b-5e113f5cd392 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.556675] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d52175ed-5e03-4462-b18d-f67a0c67a5e1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.570363] env[63088]: DEBUG nova.compute.provider_tree [None req-7af44f64-0067-49a1-ac0c-3ee5b8047867 tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 741.676669] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquiring lock "4971b24c-6710-4f50-9846-727dad264b1f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 741.676891] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Lock "4971b24c-6710-4f50-9846-727dad264b1f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 741.776011] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3edaa40f-e8ca-4f6f-a9f2-5744f490f65d tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 741.873318] env[63088]: DEBUG oslo_concurrency.lockutils [None req-ae448e13-3c80-41d5-9b1f-e650107512d4 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Releasing lock "refresh_cache-2b43620d-94ac-4c56-87df-4066f5d24088" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 741.874040] env[63088]: DEBUG nova.compute.manager [None req-ae448e13-3c80-41d5-9b1f-e650107512d4 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 741.874040] env[63088]: DEBUG nova.compute.manager [None req-ae448e13-3c80-41d5-9b1f-e650107512d4 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 741.874040] env[63088]: DEBUG nova.network.neutron [None req-ae448e13-3c80-41d5-9b1f-e650107512d4 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 741.891392] env[63088]: DEBUG nova.network.neutron [None req-ae448e13-3c80-41d5-9b1f-e650107512d4 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 742.076234] env[63088]: DEBUG nova.scheduler.client.report [None req-7af44f64-0067-49a1-ac0c-3ee5b8047867 tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 742.104026] env[63088]: INFO nova.scheduler.client.report [None req-a2a0ac90-0e7a-45d1-a070-f7ee6be01d68 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Deleted allocations for instance eb42c667-b0ae-4807-8f69-88dd671aa7d7 [ 742.394793] env[63088]: DEBUG nova.network.neutron [None req-ae448e13-3c80-41d5-9b1f-e650107512d4 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 742.583248] env[63088]: DEBUG oslo_concurrency.lockutils [None req-7af44f64-0067-49a1-ac0c-3ee5b8047867 tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.861s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 742.584632] env[63088]: ERROR nova.compute.manager [None req-7af44f64-0067-49a1-ac0c-3ee5b8047867 tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 047bc58e-8cf5-4c4d-b766-962ea1063144, please check neutron logs for more information. [ 742.584632] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] Traceback (most recent call last): [ 742.584632] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 742.584632] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] self.driver.spawn(context, instance, image_meta, [ 742.584632] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 742.584632] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 742.584632] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 742.584632] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] vm_ref = self.build_virtual_machine(instance, [ 742.584632] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 742.584632] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] vif_infos = vmwarevif.get_vif_info(self._session, [ 742.584632] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 742.585049] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] for vif in network_info: [ 742.585049] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 742.585049] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] return self._sync_wrapper(fn, *args, **kwargs) [ 742.585049] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 742.585049] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] self.wait() [ 742.585049] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 742.585049] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] self[:] = self._gt.wait() [ 742.585049] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 742.585049] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] return self._exit_event.wait() [ 742.585049] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 742.585049] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] result = hub.switch() [ 742.585049] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 742.585049] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] return self.greenlet.switch() [ 742.585457] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 742.585457] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] result = function(*args, **kwargs) [ 742.585457] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 742.585457] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] return func(*args, **kwargs) [ 742.585457] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 742.585457] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] raise e [ 742.585457] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 742.585457] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] nwinfo = self.network_api.allocate_for_instance( [ 742.585457] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 742.585457] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] created_port_ids = self._update_ports_for_instance( [ 742.585457] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 742.585457] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] with excutils.save_and_reraise_exception(): [ 742.585457] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 742.585860] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] self.force_reraise() [ 742.585860] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 742.585860] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] raise self.value [ 742.585860] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 742.585860] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] updated_port = self._update_port( [ 742.585860] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 742.585860] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] _ensure_no_port_binding_failure(port) [ 742.585860] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 742.585860] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] raise exception.PortBindingFailed(port_id=port['id']) [ 742.585860] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] nova.exception.PortBindingFailed: Binding failed for port 047bc58e-8cf5-4c4d-b766-962ea1063144, please check neutron logs for more information. [ 742.585860] env[63088]: ERROR nova.compute.manager [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] [ 742.586213] env[63088]: DEBUG nova.compute.utils [None req-7af44f64-0067-49a1-ac0c-3ee5b8047867 tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] Binding failed for port 047bc58e-8cf5-4c4d-b766-962ea1063144, please check neutron logs for more information. {{(pid=63088) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 742.586546] env[63088]: DEBUG oslo_concurrency.lockutils [None req-03263d3e-caae-4f24-95aa-ac7aedd64259 tempest-ImagesOneServerTestJSON-44065314 tempest-ImagesOneServerTestJSON-44065314-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.909s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 742.589344] env[63088]: DEBUG nova.compute.manager [None req-7af44f64-0067-49a1-ac0c-3ee5b8047867 tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] Build of instance f2b4485c-ebcb-4063-a25a-e2efe6d1758f was re-scheduled: Binding failed for port 047bc58e-8cf5-4c4d-b766-962ea1063144, please check neutron logs for more information. {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 742.589909] env[63088]: DEBUG nova.compute.manager [None req-7af44f64-0067-49a1-ac0c-3ee5b8047867 tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] Unplugging VIFs for instance {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 742.589982] env[63088]: DEBUG oslo_concurrency.lockutils [None req-7af44f64-0067-49a1-ac0c-3ee5b8047867 tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] Acquiring lock "refresh_cache-f2b4485c-ebcb-4063-a25a-e2efe6d1758f" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 742.590153] env[63088]: DEBUG oslo_concurrency.lockutils [None req-7af44f64-0067-49a1-ac0c-3ee5b8047867 tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] Acquired lock "refresh_cache-f2b4485c-ebcb-4063-a25a-e2efe6d1758f" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 742.590310] env[63088]: DEBUG nova.network.neutron [None req-7af44f64-0067-49a1-ac0c-3ee5b8047867 tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 742.615668] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a2a0ac90-0e7a-45d1-a070-f7ee6be01d68 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Lock "eb42c667-b0ae-4807-8f69-88dd671aa7d7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 143.224s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 742.900651] env[63088]: INFO nova.compute.manager [None req-ae448e13-3c80-41d5-9b1f-e650107512d4 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 2b43620d-94ac-4c56-87df-4066f5d24088] Took 1.02 seconds to deallocate network for instance. [ 743.115502] env[63088]: DEBUG nova.network.neutron [None req-7af44f64-0067-49a1-ac0c-3ee5b8047867 tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 743.120208] env[63088]: DEBUG nova.compute.manager [None req-3a58af89-5e90-4bc8-9f8f-c02b80d14311 tempest-InstanceActionsTestJSON-1717540784 tempest-InstanceActionsTestJSON-1717540784-project-member] [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 743.220583] env[63088]: DEBUG nova.network.neutron [None req-7af44f64-0067-49a1-ac0c-3ee5b8047867 tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 743.384958] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0887c362-8996-4da8-bcdb-e9eed6c739f9 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.393017] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-763371f9-5c7d-47a7-bf8f-0ee328cf4951 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.427488] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-823d9116-e1a7-48b8-8194-fa67386beca9 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.434799] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d64c271f-ee15-468e-9a72-8b319e7f238f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.447706] env[63088]: DEBUG nova.compute.provider_tree [None req-03263d3e-caae-4f24-95aa-ac7aedd64259 tempest-ImagesOneServerTestJSON-44065314 tempest-ImagesOneServerTestJSON-44065314-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 743.638903] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3a58af89-5e90-4bc8-9f8f-c02b80d14311 tempest-InstanceActionsTestJSON-1717540784 tempest-InstanceActionsTestJSON-1717540784-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 743.727256] env[63088]: DEBUG oslo_concurrency.lockutils [None req-7af44f64-0067-49a1-ac0c-3ee5b8047867 tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] Releasing lock "refresh_cache-f2b4485c-ebcb-4063-a25a-e2efe6d1758f" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 743.727513] env[63088]: DEBUG nova.compute.manager [None req-7af44f64-0067-49a1-ac0c-3ee5b8047867 tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 743.727692] env[63088]: DEBUG nova.compute.manager [None req-7af44f64-0067-49a1-ac0c-3ee5b8047867 tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 743.727861] env[63088]: DEBUG nova.network.neutron [None req-7af44f64-0067-49a1-ac0c-3ee5b8047867 tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 743.744460] env[63088]: DEBUG nova.network.neutron [None req-7af44f64-0067-49a1-ac0c-3ee5b8047867 tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 743.816641] env[63088]: DEBUG oslo_concurrency.lockutils [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquiring lock "4847f15e-98d4-401c-91ff-067e84a85727" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 743.816864] env[63088]: DEBUG oslo_concurrency.lockutils [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Lock "4847f15e-98d4-401c-91ff-067e84a85727" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 743.950429] env[63088]: INFO nova.scheduler.client.report [None req-ae448e13-3c80-41d5-9b1f-e650107512d4 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Deleted allocations for instance 2b43620d-94ac-4c56-87df-4066f5d24088 [ 743.956148] env[63088]: DEBUG nova.scheduler.client.report [None req-03263d3e-caae-4f24-95aa-ac7aedd64259 tempest-ImagesOneServerTestJSON-44065314 tempest-ImagesOneServerTestJSON-44065314-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 744.246501] env[63088]: DEBUG nova.network.neutron [None req-7af44f64-0067-49a1-ac0c-3ee5b8047867 tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 744.461935] env[63088]: DEBUG oslo_concurrency.lockutils [None req-03263d3e-caae-4f24-95aa-ac7aedd64259 tempest-ImagesOneServerTestJSON-44065314 tempest-ImagesOneServerTestJSON-44065314-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.875s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 744.462595] env[63088]: ERROR nova.compute.manager [None req-03263d3e-caae-4f24-95aa-ac7aedd64259 tempest-ImagesOneServerTestJSON-44065314 tempest-ImagesOneServerTestJSON-44065314-project-member] [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 403c6f43-6162-47c8-a025-754113fc33d6, please check neutron logs for more information. [ 744.462595] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] Traceback (most recent call last): [ 744.462595] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 744.462595] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] self.driver.spawn(context, instance, image_meta, [ 744.462595] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 744.462595] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 744.462595] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 744.462595] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] vm_ref = self.build_virtual_machine(instance, [ 744.462595] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 744.462595] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] vif_infos = vmwarevif.get_vif_info(self._session, [ 744.462595] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 744.462919] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] for vif in network_info: [ 744.462919] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 744.462919] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] return self._sync_wrapper(fn, *args, **kwargs) [ 744.462919] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 744.462919] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] self.wait() [ 744.462919] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 744.462919] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] self[:] = self._gt.wait() [ 744.462919] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 744.462919] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] return self._exit_event.wait() [ 744.462919] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 744.462919] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] current.throw(*self._exc) [ 744.462919] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 744.462919] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] result = function(*args, **kwargs) [ 744.463520] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 744.463520] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] return func(*args, **kwargs) [ 744.463520] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 744.463520] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] raise e [ 744.463520] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 744.463520] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] nwinfo = self.network_api.allocate_for_instance( [ 744.463520] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 744.463520] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] created_port_ids = self._update_ports_for_instance( [ 744.463520] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 744.463520] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] with excutils.save_and_reraise_exception(): [ 744.463520] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 744.463520] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] self.force_reraise() [ 744.463520] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 744.463929] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] raise self.value [ 744.463929] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 744.463929] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] updated_port = self._update_port( [ 744.463929] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 744.463929] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] _ensure_no_port_binding_failure(port) [ 744.463929] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 744.463929] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] raise exception.PortBindingFailed(port_id=port['id']) [ 744.463929] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] nova.exception.PortBindingFailed: Binding failed for port 403c6f43-6162-47c8-a025-754113fc33d6, please check neutron logs for more information. [ 744.463929] env[63088]: ERROR nova.compute.manager [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] [ 744.463929] env[63088]: DEBUG nova.compute.utils [None req-03263d3e-caae-4f24-95aa-ac7aedd64259 tempest-ImagesOneServerTestJSON-44065314 tempest-ImagesOneServerTestJSON-44065314-project-member] [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] Binding failed for port 403c6f43-6162-47c8-a025-754113fc33d6, please check neutron logs for more information. {{(pid=63088) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 744.464587] env[63088]: DEBUG oslo_concurrency.lockutils [None req-ae448e13-3c80-41d5-9b1f-e650107512d4 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lock "2b43620d-94ac-4c56-87df-4066f5d24088" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 143.315s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 744.465010] env[63088]: DEBUG nova.compute.manager [None req-03263d3e-caae-4f24-95aa-ac7aedd64259 tempest-ImagesOneServerTestJSON-44065314 tempest-ImagesOneServerTestJSON-44065314-project-member] [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] Build of instance c63c1740-fcec-47fa-a7b9-986b83c4abd9 was re-scheduled: Binding failed for port 403c6f43-6162-47c8-a025-754113fc33d6, please check neutron logs for more information. {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 744.465422] env[63088]: DEBUG nova.compute.manager [None req-03263d3e-caae-4f24-95aa-ac7aedd64259 tempest-ImagesOneServerTestJSON-44065314 tempest-ImagesOneServerTestJSON-44065314-project-member] [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] Unplugging VIFs for instance {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 744.465630] env[63088]: DEBUG oslo_concurrency.lockutils [None req-03263d3e-caae-4f24-95aa-ac7aedd64259 tempest-ImagesOneServerTestJSON-44065314 tempest-ImagesOneServerTestJSON-44065314-project-member] Acquiring lock "refresh_cache-c63c1740-fcec-47fa-a7b9-986b83c4abd9" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 744.465770] env[63088]: DEBUG oslo_concurrency.lockutils [None req-03263d3e-caae-4f24-95aa-ac7aedd64259 tempest-ImagesOneServerTestJSON-44065314 tempest-ImagesOneServerTestJSON-44065314-project-member] Acquired lock "refresh_cache-c63c1740-fcec-47fa-a7b9-986b83c4abd9" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 744.465925] env[63088]: DEBUG nova.network.neutron [None req-03263d3e-caae-4f24-95aa-ac7aedd64259 tempest-ImagesOneServerTestJSON-44065314 tempest-ImagesOneServerTestJSON-44065314-project-member] [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 744.466861] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d012c7de-73d3-432e-8cc9-9d2f0456777d tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.089s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 744.468349] env[63088]: INFO nova.compute.claims [None req-d012c7de-73d3-432e-8cc9-9d2f0456777d tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 744.749206] env[63088]: INFO nova.compute.manager [None req-7af44f64-0067-49a1-ac0c-3ee5b8047867 tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] [instance: f2b4485c-ebcb-4063-a25a-e2efe6d1758f] Took 1.02 seconds to deallocate network for instance. [ 744.976830] env[63088]: DEBUG nova.compute.manager [None req-87ab5d07-9ebb-429d-8640-0197960e7d1a tempest-ServersNegativeTestMultiTenantJSON-2054426678 tempest-ServersNegativeTestMultiTenantJSON-2054426678-project-member] [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 744.993474] env[63088]: DEBUG nova.network.neutron [None req-03263d3e-caae-4f24-95aa-ac7aedd64259 tempest-ImagesOneServerTestJSON-44065314 tempest-ImagesOneServerTestJSON-44065314-project-member] [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 745.089575] env[63088]: DEBUG nova.network.neutron [None req-03263d3e-caae-4f24-95aa-ac7aedd64259 tempest-ImagesOneServerTestJSON-44065314 tempest-ImagesOneServerTestJSON-44065314-project-member] [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 745.497836] env[63088]: DEBUG oslo_concurrency.lockutils [None req-87ab5d07-9ebb-429d-8640-0197960e7d1a tempest-ServersNegativeTestMultiTenantJSON-2054426678 tempest-ServersNegativeTestMultiTenantJSON-2054426678-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 745.594473] env[63088]: DEBUG oslo_concurrency.lockutils [None req-03263d3e-caae-4f24-95aa-ac7aedd64259 tempest-ImagesOneServerTestJSON-44065314 tempest-ImagesOneServerTestJSON-44065314-project-member] Releasing lock "refresh_cache-c63c1740-fcec-47fa-a7b9-986b83c4abd9" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 745.594705] env[63088]: DEBUG nova.compute.manager [None req-03263d3e-caae-4f24-95aa-ac7aedd64259 tempest-ImagesOneServerTestJSON-44065314 tempest-ImagesOneServerTestJSON-44065314-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 745.594880] env[63088]: DEBUG nova.compute.manager [None req-03263d3e-caae-4f24-95aa-ac7aedd64259 tempest-ImagesOneServerTestJSON-44065314 tempest-ImagesOneServerTestJSON-44065314-project-member] [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 745.595107] env[63088]: DEBUG nova.network.neutron [None req-03263d3e-caae-4f24-95aa-ac7aedd64259 tempest-ImagesOneServerTestJSON-44065314 tempest-ImagesOneServerTestJSON-44065314-project-member] [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 745.617403] env[63088]: DEBUG nova.network.neutron [None req-03263d3e-caae-4f24-95aa-ac7aedd64259 tempest-ImagesOneServerTestJSON-44065314 tempest-ImagesOneServerTestJSON-44065314-project-member] [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 745.714207] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f07f8e64-9b70-48b9-94c9-a37507767058 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.722391] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16782b53-bde2-4a4d-ab2b-442ea77b4aad {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.761752] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b2dc28e-e9cb-422c-9ac7-ff3dfe083292 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.769204] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce73d8b8-3b9e-4bd2-9ac5-20e005ef1d2c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.782466] env[63088]: DEBUG nova.compute.provider_tree [None req-d012c7de-73d3-432e-8cc9-9d2f0456777d tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 745.784342] env[63088]: INFO nova.scheduler.client.report [None req-7af44f64-0067-49a1-ac0c-3ee5b8047867 tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] Deleted allocations for instance f2b4485c-ebcb-4063-a25a-e2efe6d1758f [ 746.120719] env[63088]: DEBUG nova.network.neutron [None req-03263d3e-caae-4f24-95aa-ac7aedd64259 tempest-ImagesOneServerTestJSON-44065314 tempest-ImagesOneServerTestJSON-44065314-project-member] [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 746.291349] env[63088]: DEBUG nova.scheduler.client.report [None req-d012c7de-73d3-432e-8cc9-9d2f0456777d tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 746.295163] env[63088]: DEBUG oslo_concurrency.lockutils [None req-7af44f64-0067-49a1-ac0c-3ee5b8047867 tempest-ServersTestMultiNic-1970395980 tempest-ServersTestMultiNic-1970395980-project-member] Lock "f2b4485c-ebcb-4063-a25a-e2efe6d1758f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 141.596s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 746.298668] env[63088]: DEBUG oslo_concurrency.lockutils [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquiring lock "e845f60b-004f-4a5c-813e-0c183e99eba6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 746.298869] env[63088]: DEBUG oslo_concurrency.lockutils [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lock "e845f60b-004f-4a5c-813e-0c183e99eba6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 746.626916] env[63088]: INFO nova.compute.manager [None req-03263d3e-caae-4f24-95aa-ac7aedd64259 tempest-ImagesOneServerTestJSON-44065314 tempest-ImagesOneServerTestJSON-44065314-project-member] [instance: c63c1740-fcec-47fa-a7b9-986b83c4abd9] Took 1.03 seconds to deallocate network for instance. [ 746.798485] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d012c7de-73d3-432e-8cc9-9d2f0456777d tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.331s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 746.799017] env[63088]: DEBUG nova.compute.manager [None req-d012c7de-73d3-432e-8cc9-9d2f0456777d tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 746.801869] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a6077124-74e2-4777-9beb-9998c6275315 tempest-ServersTestBootFromVolume-2105176458 tempest-ServersTestBootFromVolume-2105176458-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 11.190s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 746.804685] env[63088]: DEBUG nova.compute.manager [None req-aa39901a-ab16-4a54-a2f6-147052a0ed90 tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] [instance: 43ded929-6890-4296-ae90-230f7c8b2595] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 747.308631] env[63088]: DEBUG nova.compute.utils [None req-d012c7de-73d3-432e-8cc9-9d2f0456777d tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 747.315654] env[63088]: DEBUG nova.compute.manager [None req-d012c7de-73d3-432e-8cc9-9d2f0456777d tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 747.315817] env[63088]: DEBUG nova.network.neutron [None req-d012c7de-73d3-432e-8cc9-9d2f0456777d tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 747.332069] env[63088]: DEBUG oslo_concurrency.lockutils [None req-aa39901a-ab16-4a54-a2f6-147052a0ed90 tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 747.389198] env[63088]: DEBUG nova.policy [None req-d012c7de-73d3-432e-8cc9-9d2f0456777d tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ca8c6f64e1724943aaef6e07fa3b7308', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'edf7d041b681480791590601dde8ba97', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 747.561979] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64ef02c1-48c1-4439-a641-a1745ae75a64 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.569560] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef7a7e05-3440-4f56-8b80-b2c8663a82fc {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.600199] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba0df12f-e9f0-4115-afaa-4f9bc688eefc {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.609953] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58d2641a-89a6-4fe4-9827-10f0c066408a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.624739] env[63088]: DEBUG nova.compute.provider_tree [None req-a6077124-74e2-4777-9beb-9998c6275315 tempest-ServersTestBootFromVolume-2105176458 tempest-ServersTestBootFromVolume-2105176458-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 747.661606] env[63088]: INFO nova.scheduler.client.report [None req-03263d3e-caae-4f24-95aa-ac7aedd64259 tempest-ImagesOneServerTestJSON-44065314 tempest-ImagesOneServerTestJSON-44065314-project-member] Deleted allocations for instance c63c1740-fcec-47fa-a7b9-986b83c4abd9 [ 747.818971] env[63088]: DEBUG nova.compute.manager [None req-d012c7de-73d3-432e-8cc9-9d2f0456777d tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 748.109541] env[63088]: DEBUG nova.network.neutron [None req-d012c7de-73d3-432e-8cc9-9d2f0456777d tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] Successfully created port: 170f1cb1-c3ef-431f-8ef9-de984c0f19d1 {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 748.129223] env[63088]: DEBUG nova.scheduler.client.report [None req-a6077124-74e2-4777-9beb-9998c6275315 tempest-ServersTestBootFromVolume-2105176458 tempest-ServersTestBootFromVolume-2105176458-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 748.169886] env[63088]: DEBUG oslo_concurrency.lockutils [None req-03263d3e-caae-4f24-95aa-ac7aedd64259 tempest-ImagesOneServerTestJSON-44065314 tempest-ImagesOneServerTestJSON-44065314-project-member] Lock "c63c1740-fcec-47fa-a7b9-986b83c4abd9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 142.368s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 748.634336] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a6077124-74e2-4777-9beb-9998c6275315 tempest-ServersTestBootFromVolume-2105176458 tempest-ServersTestBootFromVolume-2105176458-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.832s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 748.634961] env[63088]: ERROR nova.compute.manager [None req-a6077124-74e2-4777-9beb-9998c6275315 tempest-ServersTestBootFromVolume-2105176458 tempest-ServersTestBootFromVolume-2105176458-project-member] [instance: 407db84f-3322-42bc-baac-5528e856af8d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8193ccbb-9d49-4c2e-a0b5-7a90dba1cb62, please check neutron logs for more information. [ 748.634961] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] Traceback (most recent call last): [ 748.634961] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 748.634961] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] self.driver.spawn(context, instance, image_meta, [ 748.634961] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 748.634961] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 748.634961] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 748.634961] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] vm_ref = self.build_virtual_machine(instance, [ 748.634961] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 748.634961] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] vif_infos = vmwarevif.get_vif_info(self._session, [ 748.634961] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 748.635466] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] for vif in network_info: [ 748.635466] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 748.635466] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] return self._sync_wrapper(fn, *args, **kwargs) [ 748.635466] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 748.635466] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] self.wait() [ 748.635466] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 748.635466] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] self[:] = self._gt.wait() [ 748.635466] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 748.635466] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] return self._exit_event.wait() [ 748.635466] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 748.635466] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] current.throw(*self._exc) [ 748.635466] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 748.635466] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] result = function(*args, **kwargs) [ 748.636240] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 748.636240] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] return func(*args, **kwargs) [ 748.636240] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 748.636240] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] raise e [ 748.636240] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 748.636240] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] nwinfo = self.network_api.allocate_for_instance( [ 748.636240] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 748.636240] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] created_port_ids = self._update_ports_for_instance( [ 748.636240] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 748.636240] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] with excutils.save_and_reraise_exception(): [ 748.636240] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 748.636240] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] self.force_reraise() [ 748.636240] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 748.636844] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] raise self.value [ 748.636844] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 748.636844] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] updated_port = self._update_port( [ 748.636844] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 748.636844] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] _ensure_no_port_binding_failure(port) [ 748.636844] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 748.636844] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] raise exception.PortBindingFailed(port_id=port['id']) [ 748.636844] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] nova.exception.PortBindingFailed: Binding failed for port 8193ccbb-9d49-4c2e-a0b5-7a90dba1cb62, please check neutron logs for more information. [ 748.636844] env[63088]: ERROR nova.compute.manager [instance: 407db84f-3322-42bc-baac-5528e856af8d] [ 748.636844] env[63088]: DEBUG nova.compute.utils [None req-a6077124-74e2-4777-9beb-9998c6275315 tempest-ServersTestBootFromVolume-2105176458 tempest-ServersTestBootFromVolume-2105176458-project-member] [instance: 407db84f-3322-42bc-baac-5528e856af8d] Binding failed for port 8193ccbb-9d49-4c2e-a0b5-7a90dba1cb62, please check neutron logs for more information. {{(pid=63088) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 748.637673] env[63088]: DEBUG nova.compute.manager [None req-a6077124-74e2-4777-9beb-9998c6275315 tempest-ServersTestBootFromVolume-2105176458 tempest-ServersTestBootFromVolume-2105176458-project-member] [instance: 407db84f-3322-42bc-baac-5528e856af8d] Build of instance 407db84f-3322-42bc-baac-5528e856af8d was re-scheduled: Binding failed for port 8193ccbb-9d49-4c2e-a0b5-7a90dba1cb62, please check neutron logs for more information. {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 748.638130] env[63088]: DEBUG nova.compute.manager [None req-a6077124-74e2-4777-9beb-9998c6275315 tempest-ServersTestBootFromVolume-2105176458 tempest-ServersTestBootFromVolume-2105176458-project-member] [instance: 407db84f-3322-42bc-baac-5528e856af8d] Unplugging VIFs for instance {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 748.638367] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a6077124-74e2-4777-9beb-9998c6275315 tempest-ServersTestBootFromVolume-2105176458 tempest-ServersTestBootFromVolume-2105176458-project-member] Acquiring lock "refresh_cache-407db84f-3322-42bc-baac-5528e856af8d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 748.638514] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a6077124-74e2-4777-9beb-9998c6275315 tempest-ServersTestBootFromVolume-2105176458 tempest-ServersTestBootFromVolume-2105176458-project-member] Acquired lock "refresh_cache-407db84f-3322-42bc-baac-5528e856af8d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 748.638674] env[63088]: DEBUG nova.network.neutron [None req-a6077124-74e2-4777-9beb-9998c6275315 tempest-ServersTestBootFromVolume-2105176458 tempest-ServersTestBootFromVolume-2105176458-project-member] [instance: 407db84f-3322-42bc-baac-5528e856af8d] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 748.643019] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3b417ae5-c397-4c95-9150-090b3a6260b8 tempest-FloatingIPsAssociationTestJSON-911050413 tempest-FloatingIPsAssociationTestJSON-911050413-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.985s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 748.673049] env[63088]: DEBUG nova.compute.manager [None req-1c0c7b6c-8782-4a52-b8ee-b7fd4463c345 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 748.830296] env[63088]: DEBUG nova.compute.manager [None req-d012c7de-73d3-432e-8cc9-9d2f0456777d tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 748.860518] env[63088]: DEBUG nova.virt.hardware [None req-d012c7de-73d3-432e-8cc9-9d2f0456777d tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 748.860809] env[63088]: DEBUG nova.virt.hardware [None req-d012c7de-73d3-432e-8cc9-9d2f0456777d tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 748.860973] env[63088]: DEBUG nova.virt.hardware [None req-d012c7de-73d3-432e-8cc9-9d2f0456777d tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 748.861588] env[63088]: DEBUG nova.virt.hardware [None req-d012c7de-73d3-432e-8cc9-9d2f0456777d tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 748.861588] env[63088]: DEBUG nova.virt.hardware [None req-d012c7de-73d3-432e-8cc9-9d2f0456777d tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 748.861588] env[63088]: DEBUG nova.virt.hardware [None req-d012c7de-73d3-432e-8cc9-9d2f0456777d tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 748.862011] env[63088]: DEBUG nova.virt.hardware [None req-d012c7de-73d3-432e-8cc9-9d2f0456777d tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 748.862011] env[63088]: DEBUG nova.virt.hardware [None req-d012c7de-73d3-432e-8cc9-9d2f0456777d tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 748.862202] env[63088]: DEBUG nova.virt.hardware [None req-d012c7de-73d3-432e-8cc9-9d2f0456777d tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 748.862276] env[63088]: DEBUG nova.virt.hardware [None req-d012c7de-73d3-432e-8cc9-9d2f0456777d tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 748.862451] env[63088]: DEBUG nova.virt.hardware [None req-d012c7de-73d3-432e-8cc9-9d2f0456777d tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 748.863364] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bc5d5a6-c799-4d8d-9461-2075ec50fd2b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.872085] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ee0db5c-2e7f-4290-92cb-7e1845c6f602 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.188903] env[63088]: DEBUG nova.compute.manager [req-97b3cdb1-37f1-438f-bdfb-7449c691efb1 req-820209b9-e267-42d2-9dac-aa00b29d1ca8 service nova] [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] Received event network-changed-170f1cb1-c3ef-431f-8ef9-de984c0f19d1 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 749.189708] env[63088]: DEBUG nova.compute.manager [req-97b3cdb1-37f1-438f-bdfb-7449c691efb1 req-820209b9-e267-42d2-9dac-aa00b29d1ca8 service nova] [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] Refreshing instance network info cache due to event network-changed-170f1cb1-c3ef-431f-8ef9-de984c0f19d1. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 749.189988] env[63088]: DEBUG oslo_concurrency.lockutils [req-97b3cdb1-37f1-438f-bdfb-7449c691efb1 req-820209b9-e267-42d2-9dac-aa00b29d1ca8 service nova] Acquiring lock "refresh_cache-641eed47-54f6-4c08-a8fd-f06bc5f7fb56" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 749.190921] env[63088]: DEBUG oslo_concurrency.lockutils [req-97b3cdb1-37f1-438f-bdfb-7449c691efb1 req-820209b9-e267-42d2-9dac-aa00b29d1ca8 service nova] Acquired lock "refresh_cache-641eed47-54f6-4c08-a8fd-f06bc5f7fb56" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 749.191241] env[63088]: DEBUG nova.network.neutron [req-97b3cdb1-37f1-438f-bdfb-7449c691efb1 req-820209b9-e267-42d2-9dac-aa00b29d1ca8 service nova] [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] Refreshing network info cache for port 170f1cb1-c3ef-431f-8ef9-de984c0f19d1 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 749.202009] env[63088]: DEBUG oslo_concurrency.lockutils [None req-1c0c7b6c-8782-4a52-b8ee-b7fd4463c345 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 749.318592] env[63088]: DEBUG nova.network.neutron [None req-a6077124-74e2-4777-9beb-9998c6275315 tempest-ServersTestBootFromVolume-2105176458 tempest-ServersTestBootFromVolume-2105176458-project-member] [instance: 407db84f-3322-42bc-baac-5528e856af8d] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 749.383084] env[63088]: ERROR nova.compute.manager [None req-d012c7de-73d3-432e-8cc9-9d2f0456777d tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 170f1cb1-c3ef-431f-8ef9-de984c0f19d1, please check neutron logs for more information. [ 749.383084] env[63088]: ERROR nova.compute.manager Traceback (most recent call last): [ 749.383084] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 749.383084] env[63088]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 749.383084] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 749.383084] env[63088]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 749.383084] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 749.383084] env[63088]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 749.383084] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 749.383084] env[63088]: ERROR nova.compute.manager self.force_reraise() [ 749.383084] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 749.383084] env[63088]: ERROR nova.compute.manager raise self.value [ 749.383084] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 749.383084] env[63088]: ERROR nova.compute.manager updated_port = self._update_port( [ 749.383084] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 749.383084] env[63088]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 749.384915] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 749.384915] env[63088]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 749.384915] env[63088]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 170f1cb1-c3ef-431f-8ef9-de984c0f19d1, please check neutron logs for more information. [ 749.384915] env[63088]: ERROR nova.compute.manager [ 749.384915] env[63088]: Traceback (most recent call last): [ 749.384915] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 749.384915] env[63088]: listener.cb(fileno) [ 749.384915] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 749.384915] env[63088]: result = function(*args, **kwargs) [ 749.384915] env[63088]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 749.384915] env[63088]: return func(*args, **kwargs) [ 749.384915] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 749.384915] env[63088]: raise e [ 749.384915] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 749.384915] env[63088]: nwinfo = self.network_api.allocate_for_instance( [ 749.384915] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 749.384915] env[63088]: created_port_ids = self._update_ports_for_instance( [ 749.384915] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 749.384915] env[63088]: with excutils.save_and_reraise_exception(): [ 749.384915] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 749.384915] env[63088]: self.force_reraise() [ 749.384915] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 749.384915] env[63088]: raise self.value [ 749.384915] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 749.384915] env[63088]: updated_port = self._update_port( [ 749.384915] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 749.384915] env[63088]: _ensure_no_port_binding_failure(port) [ 749.384915] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 749.384915] env[63088]: raise exception.PortBindingFailed(port_id=port['id']) [ 749.386337] env[63088]: nova.exception.PortBindingFailed: Binding failed for port 170f1cb1-c3ef-431f-8ef9-de984c0f19d1, please check neutron logs for more information. [ 749.386337] env[63088]: Removing descriptor: 17 [ 749.386337] env[63088]: ERROR nova.compute.manager [None req-d012c7de-73d3-432e-8cc9-9d2f0456777d tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 170f1cb1-c3ef-431f-8ef9-de984c0f19d1, please check neutron logs for more information. [ 749.386337] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] Traceback (most recent call last): [ 749.386337] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 749.386337] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] yield resources [ 749.386337] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 749.386337] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] self.driver.spawn(context, instance, image_meta, [ 749.386337] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 749.386337] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] self._vmops.spawn(context, instance, image_meta, injected_files, [ 749.386337] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 749.386337] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] vm_ref = self.build_virtual_machine(instance, [ 749.386831] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 749.386831] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] vif_infos = vmwarevif.get_vif_info(self._session, [ 749.386831] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 749.386831] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] for vif in network_info: [ 749.386831] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 749.386831] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] return self._sync_wrapper(fn, *args, **kwargs) [ 749.386831] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 749.386831] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] self.wait() [ 749.386831] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 749.386831] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] self[:] = self._gt.wait() [ 749.386831] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 749.386831] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] return self._exit_event.wait() [ 749.386831] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 749.387302] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] result = hub.switch() [ 749.387302] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 749.387302] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] return self.greenlet.switch() [ 749.387302] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 749.387302] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] result = function(*args, **kwargs) [ 749.387302] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 749.387302] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] return func(*args, **kwargs) [ 749.387302] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 749.387302] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] raise e [ 749.387302] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 749.387302] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] nwinfo = self.network_api.allocate_for_instance( [ 749.387302] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 749.387302] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] created_port_ids = self._update_ports_for_instance( [ 749.387705] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 749.387705] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] with excutils.save_and_reraise_exception(): [ 749.387705] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 749.387705] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] self.force_reraise() [ 749.387705] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 749.387705] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] raise self.value [ 749.387705] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 749.387705] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] updated_port = self._update_port( [ 749.387705] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 749.387705] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] _ensure_no_port_binding_failure(port) [ 749.387705] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 749.387705] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] raise exception.PortBindingFailed(port_id=port['id']) [ 749.388121] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] nova.exception.PortBindingFailed: Binding failed for port 170f1cb1-c3ef-431f-8ef9-de984c0f19d1, please check neutron logs for more information. [ 749.388121] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] [ 749.388121] env[63088]: INFO nova.compute.manager [None req-d012c7de-73d3-432e-8cc9-9d2f0456777d tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] Terminating instance [ 749.389603] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d012c7de-73d3-432e-8cc9-9d2f0456777d tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Acquiring lock "refresh_cache-641eed47-54f6-4c08-a8fd-f06bc5f7fb56" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 749.418172] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86d02de0-feb0-4442-a9f7-5d51a32e2555 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.427025] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0fa9e6e-28ca-4c91-aa14-66d527e4f823 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.463402] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e19c111-c411-4437-8e74-f144f2d9ef9a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.469885] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9abc880f-700a-40db-889b-39c09bbaf904 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.484071] env[63088]: DEBUG nova.compute.provider_tree [None req-3b417ae5-c397-4c95-9150-090b3a6260b8 tempest-FloatingIPsAssociationTestJSON-911050413 tempest-FloatingIPsAssociationTestJSON-911050413-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 749.507715] env[63088]: DEBUG nova.network.neutron [None req-a6077124-74e2-4777-9beb-9998c6275315 tempest-ServersTestBootFromVolume-2105176458 tempest-ServersTestBootFromVolume-2105176458-project-member] [instance: 407db84f-3322-42bc-baac-5528e856af8d] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 749.722143] env[63088]: DEBUG nova.network.neutron [req-97b3cdb1-37f1-438f-bdfb-7449c691efb1 req-820209b9-e267-42d2-9dac-aa00b29d1ca8 service nova] [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 749.898402] env[63088]: DEBUG nova.network.neutron [req-97b3cdb1-37f1-438f-bdfb-7449c691efb1 req-820209b9-e267-42d2-9dac-aa00b29d1ca8 service nova] [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 749.988606] env[63088]: DEBUG nova.scheduler.client.report [None req-3b417ae5-c397-4c95-9150-090b3a6260b8 tempest-FloatingIPsAssociationTestJSON-911050413 tempest-FloatingIPsAssociationTestJSON-911050413-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 750.009663] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a6077124-74e2-4777-9beb-9998c6275315 tempest-ServersTestBootFromVolume-2105176458 tempest-ServersTestBootFromVolume-2105176458-project-member] Releasing lock "refresh_cache-407db84f-3322-42bc-baac-5528e856af8d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 750.009905] env[63088]: DEBUG nova.compute.manager [None req-a6077124-74e2-4777-9beb-9998c6275315 tempest-ServersTestBootFromVolume-2105176458 tempest-ServersTestBootFromVolume-2105176458-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 750.010095] env[63088]: DEBUG nova.compute.manager [None req-a6077124-74e2-4777-9beb-9998c6275315 tempest-ServersTestBootFromVolume-2105176458 tempest-ServersTestBootFromVolume-2105176458-project-member] [instance: 407db84f-3322-42bc-baac-5528e856af8d] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 750.010293] env[63088]: DEBUG nova.network.neutron [None req-a6077124-74e2-4777-9beb-9998c6275315 tempest-ServersTestBootFromVolume-2105176458 tempest-ServersTestBootFromVolume-2105176458-project-member] [instance: 407db84f-3322-42bc-baac-5528e856af8d] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 750.031970] env[63088]: DEBUG nova.network.neutron [None req-a6077124-74e2-4777-9beb-9998c6275315 tempest-ServersTestBootFromVolume-2105176458 tempest-ServersTestBootFromVolume-2105176458-project-member] [instance: 407db84f-3322-42bc-baac-5528e856af8d] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 750.402112] env[63088]: DEBUG oslo_concurrency.lockutils [req-97b3cdb1-37f1-438f-bdfb-7449c691efb1 req-820209b9-e267-42d2-9dac-aa00b29d1ca8 service nova] Releasing lock "refresh_cache-641eed47-54f6-4c08-a8fd-f06bc5f7fb56" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 750.402112] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d012c7de-73d3-432e-8cc9-9d2f0456777d tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Acquired lock "refresh_cache-641eed47-54f6-4c08-a8fd-f06bc5f7fb56" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 750.402216] env[63088]: DEBUG nova.network.neutron [None req-d012c7de-73d3-432e-8cc9-9d2f0456777d tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 750.496145] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3b417ae5-c397-4c95-9150-090b3a6260b8 tempest-FloatingIPsAssociationTestJSON-911050413 tempest-FloatingIPsAssociationTestJSON-911050413-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.853s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 750.496145] env[63088]: ERROR nova.compute.manager [None req-3b417ae5-c397-4c95-9150-090b3a6260b8 tempest-FloatingIPsAssociationTestJSON-911050413 tempest-FloatingIPsAssociationTestJSON-911050413-project-member] [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port eeaacdbe-07d9-4cc2-8775-873e396bf287, please check neutron logs for more information. [ 750.496145] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] Traceback (most recent call last): [ 750.496145] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 750.496145] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] self.driver.spawn(context, instance, image_meta, [ 750.496145] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 750.496145] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] self._vmops.spawn(context, instance, image_meta, injected_files, [ 750.496145] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 750.496145] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] vm_ref = self.build_virtual_machine(instance, [ 750.496424] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 750.496424] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] vif_infos = vmwarevif.get_vif_info(self._session, [ 750.496424] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 750.496424] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] for vif in network_info: [ 750.496424] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 750.496424] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] return self._sync_wrapper(fn, *args, **kwargs) [ 750.496424] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 750.496424] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] self.wait() [ 750.496424] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 750.496424] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] self[:] = self._gt.wait() [ 750.496424] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 750.496424] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] return self._exit_event.wait() [ 750.496424] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 750.496812] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] current.throw(*self._exc) [ 750.496812] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 750.496812] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] result = function(*args, **kwargs) [ 750.496812] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 750.496812] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] return func(*args, **kwargs) [ 750.496812] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 750.496812] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] raise e [ 750.496812] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 750.496812] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] nwinfo = self.network_api.allocate_for_instance( [ 750.496812] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 750.496812] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] created_port_ids = self._update_ports_for_instance( [ 750.496812] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 750.496812] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] with excutils.save_and_reraise_exception(): [ 750.497196] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 750.497196] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] self.force_reraise() [ 750.497196] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 750.497196] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] raise self.value [ 750.497196] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 750.497196] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] updated_port = self._update_port( [ 750.497196] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 750.497196] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] _ensure_no_port_binding_failure(port) [ 750.497196] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 750.497196] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] raise exception.PortBindingFailed(port_id=port['id']) [ 750.497196] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] nova.exception.PortBindingFailed: Binding failed for port eeaacdbe-07d9-4cc2-8775-873e396bf287, please check neutron logs for more information. [ 750.497196] env[63088]: ERROR nova.compute.manager [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] [ 750.497524] env[63088]: DEBUG nova.compute.utils [None req-3b417ae5-c397-4c95-9150-090b3a6260b8 tempest-FloatingIPsAssociationTestJSON-911050413 tempest-FloatingIPsAssociationTestJSON-911050413-project-member] [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] Binding failed for port eeaacdbe-07d9-4cc2-8775-873e396bf287, please check neutron logs for more information. {{(pid=63088) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 750.498226] env[63088]: DEBUG nova.compute.manager [None req-3b417ae5-c397-4c95-9150-090b3a6260b8 tempest-FloatingIPsAssociationTestJSON-911050413 tempest-FloatingIPsAssociationTestJSON-911050413-project-member] [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] Build of instance 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed was re-scheduled: Binding failed for port eeaacdbe-07d9-4cc2-8775-873e396bf287, please check neutron logs for more information. {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 750.498672] env[63088]: DEBUG nova.compute.manager [None req-3b417ae5-c397-4c95-9150-090b3a6260b8 tempest-FloatingIPsAssociationTestJSON-911050413 tempest-FloatingIPsAssociationTestJSON-911050413-project-member] [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] Unplugging VIFs for instance {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 750.498926] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3b417ae5-c397-4c95-9150-090b3a6260b8 tempest-FloatingIPsAssociationTestJSON-911050413 tempest-FloatingIPsAssociationTestJSON-911050413-project-member] Acquiring lock "refresh_cache-1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 750.499148] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3b417ae5-c397-4c95-9150-090b3a6260b8 tempest-FloatingIPsAssociationTestJSON-911050413 tempest-FloatingIPsAssociationTestJSON-911050413-project-member] Acquired lock "refresh_cache-1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 750.499400] env[63088]: DEBUG nova.network.neutron [None req-3b417ae5-c397-4c95-9150-090b3a6260b8 tempest-FloatingIPsAssociationTestJSON-911050413 tempest-FloatingIPsAssociationTestJSON-911050413-project-member] [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 750.500529] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bce00a2e-7fb0-44a9-a66e-261b7c3d5384 tempest-ServerActionsV293TestJSON-1088890325 tempest-ServerActionsV293TestJSON-1088890325-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.546s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 750.505761] env[63088]: INFO nova.compute.claims [None req-bce00a2e-7fb0-44a9-a66e-261b7c3d5384 tempest-ServerActionsV293TestJSON-1088890325 tempest-ServerActionsV293TestJSON-1088890325-project-member] [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 750.531816] env[63088]: DEBUG nova.network.neutron [None req-3b417ae5-c397-4c95-9150-090b3a6260b8 tempest-FloatingIPsAssociationTestJSON-911050413 tempest-FloatingIPsAssociationTestJSON-911050413-project-member] [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 750.533953] env[63088]: DEBUG nova.network.neutron [None req-a6077124-74e2-4777-9beb-9998c6275315 tempest-ServersTestBootFromVolume-2105176458 tempest-ServersTestBootFromVolume-2105176458-project-member] [instance: 407db84f-3322-42bc-baac-5528e856af8d] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 750.627619] env[63088]: DEBUG nova.network.neutron [None req-3b417ae5-c397-4c95-9150-090b3a6260b8 tempest-FloatingIPsAssociationTestJSON-911050413 tempest-FloatingIPsAssociationTestJSON-911050413-project-member] [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 750.927571] env[63088]: DEBUG nova.network.neutron [None req-d012c7de-73d3-432e-8cc9-9d2f0456777d tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 751.039032] env[63088]: DEBUG nova.network.neutron [None req-d012c7de-73d3-432e-8cc9-9d2f0456777d tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 751.043922] env[63088]: INFO nova.compute.manager [None req-a6077124-74e2-4777-9beb-9998c6275315 tempest-ServersTestBootFromVolume-2105176458 tempest-ServersTestBootFromVolume-2105176458-project-member] [instance: 407db84f-3322-42bc-baac-5528e856af8d] Took 1.03 seconds to deallocate network for instance. [ 751.134648] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3b417ae5-c397-4c95-9150-090b3a6260b8 tempest-FloatingIPsAssociationTestJSON-911050413 tempest-FloatingIPsAssociationTestJSON-911050413-project-member] Releasing lock "refresh_cache-1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 751.134648] env[63088]: DEBUG nova.compute.manager [None req-3b417ae5-c397-4c95-9150-090b3a6260b8 tempest-FloatingIPsAssociationTestJSON-911050413 tempest-FloatingIPsAssociationTestJSON-911050413-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 751.134648] env[63088]: DEBUG nova.compute.manager [None req-3b417ae5-c397-4c95-9150-090b3a6260b8 tempest-FloatingIPsAssociationTestJSON-911050413 tempest-FloatingIPsAssociationTestJSON-911050413-project-member] [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 751.134648] env[63088]: DEBUG nova.network.neutron [None req-3b417ae5-c397-4c95-9150-090b3a6260b8 tempest-FloatingIPsAssociationTestJSON-911050413 tempest-FloatingIPsAssociationTestJSON-911050413-project-member] [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 751.156127] env[63088]: DEBUG nova.network.neutron [None req-3b417ae5-c397-4c95-9150-090b3a6260b8 tempest-FloatingIPsAssociationTestJSON-911050413 tempest-FloatingIPsAssociationTestJSON-911050413-project-member] [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 751.216615] env[63088]: DEBUG nova.compute.manager [req-d3e057ac-cead-4a1d-96d2-52c2e2b4d15f req-d2538c39-c6c7-4c64-bdf3-e77f205bfcc3 service nova] [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] Received event network-vif-deleted-170f1cb1-c3ef-431f-8ef9-de984c0f19d1 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 751.541690] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d012c7de-73d3-432e-8cc9-9d2f0456777d tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Releasing lock "refresh_cache-641eed47-54f6-4c08-a8fd-f06bc5f7fb56" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 751.542148] env[63088]: DEBUG nova.compute.manager [None req-d012c7de-73d3-432e-8cc9-9d2f0456777d tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 751.542421] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-d012c7de-73d3-432e-8cc9-9d2f0456777d tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 751.542692] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-350faa97-b27a-45d1-bb50-c197f099e94f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.561230] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8bc0dc0-ebcf-40ac-bc7b-a40ddd9f7eda {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.585422] env[63088]: WARNING nova.virt.vmwareapi.vmops [None req-d012c7de-73d3-432e-8cc9-9d2f0456777d tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 641eed47-54f6-4c08-a8fd-f06bc5f7fb56 could not be found. [ 751.585686] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-d012c7de-73d3-432e-8cc9-9d2f0456777d tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 751.585841] env[63088]: INFO nova.compute.manager [None req-d012c7de-73d3-432e-8cc9-9d2f0456777d tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] Took 0.04 seconds to destroy the instance on the hypervisor. [ 751.586120] env[63088]: DEBUG oslo.service.loopingcall [None req-d012c7de-73d3-432e-8cc9-9d2f0456777d tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 751.589764] env[63088]: DEBUG nova.compute.manager [-] [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 751.589764] env[63088]: DEBUG nova.network.neutron [-] [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 751.605269] env[63088]: DEBUG nova.network.neutron [-] [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 751.657490] env[63088]: DEBUG nova.network.neutron [None req-3b417ae5-c397-4c95-9150-090b3a6260b8 tempest-FloatingIPsAssociationTestJSON-911050413 tempest-FloatingIPsAssociationTestJSON-911050413-project-member] [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 751.807129] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ff320bd-bee4-4fee-bb0b-1c351fcd05e3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.816974] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6926961-b511-4cd4-863f-8b4df549a569 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.852300] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc91e7ed-2ca6-4d5d-91aa-a238370d6ea6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.858608] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-068767ed-7bf9-4264-85aa-e8187960a079 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.873901] env[63088]: DEBUG nova.compute.provider_tree [None req-bce00a2e-7fb0-44a9-a66e-261b7c3d5384 tempest-ServerActionsV293TestJSON-1088890325 tempest-ServerActionsV293TestJSON-1088890325-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 752.079750] env[63088]: INFO nova.scheduler.client.report [None req-a6077124-74e2-4777-9beb-9998c6275315 tempest-ServersTestBootFromVolume-2105176458 tempest-ServersTestBootFromVolume-2105176458-project-member] Deleted allocations for instance 407db84f-3322-42bc-baac-5528e856af8d [ 752.110793] env[63088]: DEBUG nova.network.neutron [-] [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 752.162587] env[63088]: INFO nova.compute.manager [None req-3b417ae5-c397-4c95-9150-090b3a6260b8 tempest-FloatingIPsAssociationTestJSON-911050413 tempest-FloatingIPsAssociationTestJSON-911050413-project-member] [instance: 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed] Took 1.03 seconds to deallocate network for instance. [ 752.377655] env[63088]: DEBUG nova.scheduler.client.report [None req-bce00a2e-7fb0-44a9-a66e-261b7c3d5384 tempest-ServerActionsV293TestJSON-1088890325 tempest-ServerActionsV293TestJSON-1088890325-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 752.595529] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a6077124-74e2-4777-9beb-9998c6275315 tempest-ServersTestBootFromVolume-2105176458 tempest-ServersTestBootFromVolume-2105176458-project-member] Lock "407db84f-3322-42bc-baac-5528e856af8d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 132.230s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 752.614032] env[63088]: INFO nova.compute.manager [-] [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] Took 1.02 seconds to deallocate network for instance. [ 752.617599] env[63088]: DEBUG nova.compute.claims [None req-d012c7de-73d3-432e-8cc9-9d2f0456777d tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] Aborting claim: {{(pid=63088) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 752.617956] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d012c7de-73d3-432e-8cc9-9d2f0456777d tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 752.890774] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bce00a2e-7fb0-44a9-a66e-261b7c3d5384 tempest-ServerActionsV293TestJSON-1088890325 tempest-ServerActionsV293TestJSON-1088890325-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.389s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 752.890922] env[63088]: DEBUG nova.compute.manager [None req-bce00a2e-7fb0-44a9-a66e-261b7c3d5384 tempest-ServerActionsV293TestJSON-1088890325 tempest-ServerActionsV293TestJSON-1088890325-project-member] [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 752.895061] env[63088]: DEBUG oslo_concurrency.lockutils [None req-1874938e-086f-4c3c-9049-55ea0246719c tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.977s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 752.897607] env[63088]: INFO nova.compute.claims [None req-1874938e-086f-4c3c-9049-55ea0246719c tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 753.099310] env[63088]: DEBUG nova.compute.manager [None req-96c00a75-0383-4506-a4c6-d540aa18aaef tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 753.207233] env[63088]: INFO nova.scheduler.client.report [None req-3b417ae5-c397-4c95-9150-090b3a6260b8 tempest-FloatingIPsAssociationTestJSON-911050413 tempest-FloatingIPsAssociationTestJSON-911050413-project-member] Deleted allocations for instance 1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed [ 753.402730] env[63088]: DEBUG nova.compute.utils [None req-bce00a2e-7fb0-44a9-a66e-261b7c3d5384 tempest-ServerActionsV293TestJSON-1088890325 tempest-ServerActionsV293TestJSON-1088890325-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 753.406331] env[63088]: DEBUG nova.compute.manager [None req-bce00a2e-7fb0-44a9-a66e-261b7c3d5384 tempest-ServerActionsV293TestJSON-1088890325 tempest-ServerActionsV293TestJSON-1088890325-project-member] [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 753.406509] env[63088]: DEBUG nova.network.neutron [None req-bce00a2e-7fb0-44a9-a66e-261b7c3d5384 tempest-ServerActionsV293TestJSON-1088890325 tempest-ServerActionsV293TestJSON-1088890325-project-member] [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 753.473307] env[63088]: DEBUG nova.policy [None req-bce00a2e-7fb0-44a9-a66e-261b7c3d5384 tempest-ServerActionsV293TestJSON-1088890325 tempest-ServerActionsV293TestJSON-1088890325-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4b6a5b771c3b4cc9bbb6e1bb171f152f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2cca48f40ae544cb93908fa35b2fad61', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 753.629232] env[63088]: DEBUG oslo_concurrency.lockutils [None req-96c00a75-0383-4506-a4c6-d540aa18aaef tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 753.720555] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3b417ae5-c397-4c95-9150-090b3a6260b8 tempest-FloatingIPsAssociationTestJSON-911050413 tempest-FloatingIPsAssociationTestJSON-911050413-project-member] Lock "1efd14a0-bfc3-42fe-b7f1-79ad3b75f1ed" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 128.410s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 753.908661] env[63088]: DEBUG nova.compute.manager [None req-bce00a2e-7fb0-44a9-a66e-261b7c3d5384 tempest-ServerActionsV293TestJSON-1088890325 tempest-ServerActionsV293TestJSON-1088890325-project-member] [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 754.109998] env[63088]: DEBUG nova.network.neutron [None req-bce00a2e-7fb0-44a9-a66e-261b7c3d5384 tempest-ServerActionsV293TestJSON-1088890325 tempest-ServerActionsV293TestJSON-1088890325-project-member] [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] Successfully created port: 685b39f8-6e13-43dc-9dfa-6be110e1ac12 {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 754.225635] env[63088]: DEBUG nova.compute.manager [None req-02aa52fc-420f-46cd-9829-a68800c429e2 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 754.234946] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b61b6421-e0c7-4721-a8ab-e50c6d32b072 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.243189] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f69d1038-52e1-48ff-82bf-ede41577cdf8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.276844] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71be1ffc-145d-47f1-a6fa-9dee89e1da60 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.288104] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edc64a87-4e4d-4674-bc2d-43f5e2a231d9 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.303719] env[63088]: DEBUG nova.compute.provider_tree [None req-1874938e-086f-4c3c-9049-55ea0246719c tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 754.422306] env[63088]: INFO nova.virt.block_device [None req-bce00a2e-7fb0-44a9-a66e-261b7c3d5384 tempest-ServerActionsV293TestJSON-1088890325 tempest-ServerActionsV293TestJSON-1088890325-project-member] [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] Booting with volume 71116ae7-0172-4886-812a-08c051efd24c at /dev/sda [ 754.462669] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fe120e22-b9b2-4d11-99b8-d98a1c18e95e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.472641] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-745d93ce-140a-44e3-ba7d-6cdbfe2be066 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.497278] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-83d82396-a972-40aa-b549-7fd094e20600 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.506753] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cf11b68-7cef-4930-9c8a-d8fab8637ac0 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.536173] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5eda905-a5a4-4603-8a13-0e5ec7ff6bf6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.548430] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81c2011a-4e01-4a18-908c-5f2737fd2762 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.562936] env[63088]: DEBUG nova.virt.block_device [None req-bce00a2e-7fb0-44a9-a66e-261b7c3d5384 tempest-ServerActionsV293TestJSON-1088890325 tempest-ServerActionsV293TestJSON-1088890325-project-member] [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] Updating existing volume attachment record: 80685cef-58e3-41d7-8d4d-6f492eed570e {{(pid=63088) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 754.729213] env[63088]: DEBUG oslo_concurrency.lockutils [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Acquiring lock "4f6e2d99-7300-42e6-8abf-362ddd4e8c6d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 754.729213] env[63088]: DEBUG oslo_concurrency.lockutils [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Lock "4f6e2d99-7300-42e6-8abf-362ddd4e8c6d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 754.754829] env[63088]: DEBUG oslo_concurrency.lockutils [None req-02aa52fc-420f-46cd-9829-a68800c429e2 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 754.809016] env[63088]: DEBUG nova.scheduler.client.report [None req-1874938e-086f-4c3c-9049-55ea0246719c tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 755.313335] env[63088]: DEBUG oslo_concurrency.lockutils [None req-1874938e-086f-4c3c-9049-55ea0246719c tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.418s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 755.313893] env[63088]: DEBUG nova.compute.manager [None req-1874938e-086f-4c3c-9049-55ea0246719c tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 755.318686] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3edaa40f-e8ca-4f6f-a9f2-5744f490f65d tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.542s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 755.320179] env[63088]: INFO nova.compute.claims [None req-3edaa40f-e8ca-4f6f-a9f2-5744f490f65d tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 755.505586] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Acquiring lock "07fd3e6f-e0af-467c-9039-238be5c58d25" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 755.506096] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Lock "07fd3e6f-e0af-467c-9039-238be5c58d25" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 755.594391] env[63088]: DEBUG nova.compute.manager [req-85217733-6d51-4caf-9992-5ce1e729e470 req-5480de1c-2e94-4df7-9e1e-9e03fc90d9fc service nova] [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] Received event network-changed-685b39f8-6e13-43dc-9dfa-6be110e1ac12 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 755.594595] env[63088]: DEBUG nova.compute.manager [req-85217733-6d51-4caf-9992-5ce1e729e470 req-5480de1c-2e94-4df7-9e1e-9e03fc90d9fc service nova] [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] Refreshing instance network info cache due to event network-changed-685b39f8-6e13-43dc-9dfa-6be110e1ac12. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 755.594800] env[63088]: DEBUG oslo_concurrency.lockutils [req-85217733-6d51-4caf-9992-5ce1e729e470 req-5480de1c-2e94-4df7-9e1e-9e03fc90d9fc service nova] Acquiring lock "refresh_cache-b82c89ad-f7d6-4c04-a48e-acdab29ea5e7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 755.595132] env[63088]: DEBUG oslo_concurrency.lockutils [req-85217733-6d51-4caf-9992-5ce1e729e470 req-5480de1c-2e94-4df7-9e1e-9e03fc90d9fc service nova] Acquired lock "refresh_cache-b82c89ad-f7d6-4c04-a48e-acdab29ea5e7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 755.595736] env[63088]: DEBUG nova.network.neutron [req-85217733-6d51-4caf-9992-5ce1e729e470 req-5480de1c-2e94-4df7-9e1e-9e03fc90d9fc service nova] [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] Refreshing network info cache for port 685b39f8-6e13-43dc-9dfa-6be110e1ac12 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 755.828508] env[63088]: DEBUG nova.compute.utils [None req-1874938e-086f-4c3c-9049-55ea0246719c tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 755.831145] env[63088]: DEBUG nova.compute.manager [None req-1874938e-086f-4c3c-9049-55ea0246719c tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 755.834152] env[63088]: DEBUG nova.network.neutron [None req-1874938e-086f-4c3c-9049-55ea0246719c tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 755.875627] env[63088]: ERROR nova.compute.manager [None req-bce00a2e-7fb0-44a9-a66e-261b7c3d5384 tempest-ServerActionsV293TestJSON-1088890325 tempest-ServerActionsV293TestJSON-1088890325-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 685b39f8-6e13-43dc-9dfa-6be110e1ac12, please check neutron logs for more information. [ 755.875627] env[63088]: ERROR nova.compute.manager Traceback (most recent call last): [ 755.875627] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 755.875627] env[63088]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 755.875627] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 755.875627] env[63088]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 755.875627] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 755.875627] env[63088]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 755.875627] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 755.875627] env[63088]: ERROR nova.compute.manager self.force_reraise() [ 755.875627] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 755.875627] env[63088]: ERROR nova.compute.manager raise self.value [ 755.875627] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 755.875627] env[63088]: ERROR nova.compute.manager updated_port = self._update_port( [ 755.875627] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 755.875627] env[63088]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 755.876158] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 755.876158] env[63088]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 755.876158] env[63088]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 685b39f8-6e13-43dc-9dfa-6be110e1ac12, please check neutron logs for more information. [ 755.876158] env[63088]: ERROR nova.compute.manager [ 755.876158] env[63088]: Traceback (most recent call last): [ 755.876158] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 755.876158] env[63088]: listener.cb(fileno) [ 755.876158] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 755.876158] env[63088]: result = function(*args, **kwargs) [ 755.876158] env[63088]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 755.876158] env[63088]: return func(*args, **kwargs) [ 755.876158] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 755.876158] env[63088]: raise e [ 755.876158] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 755.876158] env[63088]: nwinfo = self.network_api.allocate_for_instance( [ 755.876158] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 755.876158] env[63088]: created_port_ids = self._update_ports_for_instance( [ 755.876158] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 755.876158] env[63088]: with excutils.save_and_reraise_exception(): [ 755.876158] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 755.876158] env[63088]: self.force_reraise() [ 755.876158] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 755.876158] env[63088]: raise self.value [ 755.876158] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 755.876158] env[63088]: updated_port = self._update_port( [ 755.876158] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 755.876158] env[63088]: _ensure_no_port_binding_failure(port) [ 755.876158] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 755.876158] env[63088]: raise exception.PortBindingFailed(port_id=port['id']) [ 755.877080] env[63088]: nova.exception.PortBindingFailed: Binding failed for port 685b39f8-6e13-43dc-9dfa-6be110e1ac12, please check neutron logs for more information. [ 755.877080] env[63088]: Removing descriptor: 15 [ 755.896921] env[63088]: DEBUG nova.policy [None req-1874938e-086f-4c3c-9049-55ea0246719c tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c54bbb9377894ca4bfd98d74dab12c6f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '18e424c0af324c33ad0b94a268c4651d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 756.130046] env[63088]: DEBUG nova.network.neutron [req-85217733-6d51-4caf-9992-5ce1e729e470 req-5480de1c-2e94-4df7-9e1e-9e03fc90d9fc service nova] [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 756.300791] env[63088]: DEBUG nova.network.neutron [None req-1874938e-086f-4c3c-9049-55ea0246719c tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] Successfully created port: 63967be2-f880-4637-9e38-8b244274427d {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 756.309239] env[63088]: DEBUG nova.network.neutron [req-85217733-6d51-4caf-9992-5ce1e729e470 req-5480de1c-2e94-4df7-9e1e-9e03fc90d9fc service nova] [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 756.337692] env[63088]: DEBUG nova.compute.manager [None req-1874938e-086f-4c3c-9049-55ea0246719c tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 756.668200] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3046af1c-0871-4b50-94d7-6c025ec0762d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.675841] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ea756db-afdf-4419-af76-75e36b2d5055 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.707830] env[63088]: DEBUG nova.compute.manager [None req-bce00a2e-7fb0-44a9-a66e-261b7c3d5384 tempest-ServerActionsV293TestJSON-1088890325 tempest-ServerActionsV293TestJSON-1088890325-project-member] [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 756.711392] env[63088]: DEBUG nova.virt.hardware [None req-bce00a2e-7fb0-44a9-a66e-261b7c3d5384 tempest-ServerActionsV293TestJSON-1088890325 tempest-ServerActionsV293TestJSON-1088890325-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 756.711392] env[63088]: DEBUG nova.virt.hardware [None req-bce00a2e-7fb0-44a9-a66e-261b7c3d5384 tempest-ServerActionsV293TestJSON-1088890325 tempest-ServerActionsV293TestJSON-1088890325-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 756.711392] env[63088]: DEBUG nova.virt.hardware [None req-bce00a2e-7fb0-44a9-a66e-261b7c3d5384 tempest-ServerActionsV293TestJSON-1088890325 tempest-ServerActionsV293TestJSON-1088890325-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 756.711392] env[63088]: DEBUG nova.virt.hardware [None req-bce00a2e-7fb0-44a9-a66e-261b7c3d5384 tempest-ServerActionsV293TestJSON-1088890325 tempest-ServerActionsV293TestJSON-1088890325-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 756.711645] env[63088]: DEBUG nova.virt.hardware [None req-bce00a2e-7fb0-44a9-a66e-261b7c3d5384 tempest-ServerActionsV293TestJSON-1088890325 tempest-ServerActionsV293TestJSON-1088890325-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 756.711645] env[63088]: DEBUG nova.virt.hardware [None req-bce00a2e-7fb0-44a9-a66e-261b7c3d5384 tempest-ServerActionsV293TestJSON-1088890325 tempest-ServerActionsV293TestJSON-1088890325-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 756.711844] env[63088]: DEBUG nova.virt.hardware [None req-bce00a2e-7fb0-44a9-a66e-261b7c3d5384 tempest-ServerActionsV293TestJSON-1088890325 tempest-ServerActionsV293TestJSON-1088890325-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 756.711844] env[63088]: DEBUG nova.virt.hardware [None req-bce00a2e-7fb0-44a9-a66e-261b7c3d5384 tempest-ServerActionsV293TestJSON-1088890325 tempest-ServerActionsV293TestJSON-1088890325-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 756.712012] env[63088]: DEBUG nova.virt.hardware [None req-bce00a2e-7fb0-44a9-a66e-261b7c3d5384 tempest-ServerActionsV293TestJSON-1088890325 tempest-ServerActionsV293TestJSON-1088890325-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 756.712184] env[63088]: DEBUG nova.virt.hardware [None req-bce00a2e-7fb0-44a9-a66e-261b7c3d5384 tempest-ServerActionsV293TestJSON-1088890325 tempest-ServerActionsV293TestJSON-1088890325-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 756.712376] env[63088]: DEBUG nova.virt.hardware [None req-bce00a2e-7fb0-44a9-a66e-261b7c3d5384 tempest-ServerActionsV293TestJSON-1088890325 tempest-ServerActionsV293TestJSON-1088890325-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 756.713386] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7ac088c-0754-404e-b1ad-bbec49ec2f94 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.718581] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91ad3091-6a74-4c22-9cee-70407ed77e83 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.728975] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2702dfa1-e0fb-4eb8-a0c6-20c4bc8223bd {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.734298] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-542faccb-9ac9-40cc-b1b4-532353ed20b1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.748393] env[63088]: ERROR nova.compute.manager [None req-bce00a2e-7fb0-44a9-a66e-261b7c3d5384 tempest-ServerActionsV293TestJSON-1088890325 tempest-ServerActionsV293TestJSON-1088890325-project-member] [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 685b39f8-6e13-43dc-9dfa-6be110e1ac12, please check neutron logs for more information. [ 756.748393] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] Traceback (most recent call last): [ 756.748393] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 756.748393] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] yield resources [ 756.748393] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 756.748393] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] self.driver.spawn(context, instance, image_meta, [ 756.748393] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 756.748393] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 756.748393] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 756.748393] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] vm_ref = self.build_virtual_machine(instance, [ 756.748393] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 756.748830] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] vif_infos = vmwarevif.get_vif_info(self._session, [ 756.748830] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 756.748830] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] for vif in network_info: [ 756.748830] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 756.748830] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] return self._sync_wrapper(fn, *args, **kwargs) [ 756.748830] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 756.748830] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] self.wait() [ 756.748830] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 756.748830] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] self[:] = self._gt.wait() [ 756.748830] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 756.748830] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] return self._exit_event.wait() [ 756.748830] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 756.748830] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] current.throw(*self._exc) [ 756.749238] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 756.749238] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] result = function(*args, **kwargs) [ 756.749238] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 756.749238] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] return func(*args, **kwargs) [ 756.749238] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 756.749238] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] raise e [ 756.749238] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 756.749238] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] nwinfo = self.network_api.allocate_for_instance( [ 756.749238] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 756.749238] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] created_port_ids = self._update_ports_for_instance( [ 756.749238] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 756.749238] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] with excutils.save_and_reraise_exception(): [ 756.749238] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 756.749616] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] self.force_reraise() [ 756.749616] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 756.749616] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] raise self.value [ 756.749616] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 756.749616] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] updated_port = self._update_port( [ 756.749616] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 756.749616] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] _ensure_no_port_binding_failure(port) [ 756.749616] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 756.749616] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] raise exception.PortBindingFailed(port_id=port['id']) [ 756.749616] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] nova.exception.PortBindingFailed: Binding failed for port 685b39f8-6e13-43dc-9dfa-6be110e1ac12, please check neutron logs for more information. [ 756.749616] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] [ 756.749616] env[63088]: INFO nova.compute.manager [None req-bce00a2e-7fb0-44a9-a66e-261b7c3d5384 tempest-ServerActionsV293TestJSON-1088890325 tempest-ServerActionsV293TestJSON-1088890325-project-member] [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] Terminating instance [ 756.759168] env[63088]: DEBUG nova.compute.provider_tree [None req-3edaa40f-e8ca-4f6f-a9f2-5744f490f65d tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 756.760546] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bce00a2e-7fb0-44a9-a66e-261b7c3d5384 tempest-ServerActionsV293TestJSON-1088890325 tempest-ServerActionsV293TestJSON-1088890325-project-member] Acquiring lock "refresh_cache-b82c89ad-f7d6-4c04-a48e-acdab29ea5e7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 756.811615] env[63088]: DEBUG oslo_concurrency.lockutils [req-85217733-6d51-4caf-9992-5ce1e729e470 req-5480de1c-2e94-4df7-9e1e-9e03fc90d9fc service nova] Releasing lock "refresh_cache-b82c89ad-f7d6-4c04-a48e-acdab29ea5e7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 756.812100] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bce00a2e-7fb0-44a9-a66e-261b7c3d5384 tempest-ServerActionsV293TestJSON-1088890325 tempest-ServerActionsV293TestJSON-1088890325-project-member] Acquired lock "refresh_cache-b82c89ad-f7d6-4c04-a48e-acdab29ea5e7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 756.812283] env[63088]: DEBUG nova.network.neutron [None req-bce00a2e-7fb0-44a9-a66e-261b7c3d5384 tempest-ServerActionsV293TestJSON-1088890325 tempest-ServerActionsV293TestJSON-1088890325-project-member] [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 757.265021] env[63088]: DEBUG nova.scheduler.client.report [None req-3edaa40f-e8ca-4f6f-a9f2-5744f490f65d tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 757.343470] env[63088]: DEBUG nova.network.neutron [None req-bce00a2e-7fb0-44a9-a66e-261b7c3d5384 tempest-ServerActionsV293TestJSON-1088890325 tempest-ServerActionsV293TestJSON-1088890325-project-member] [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 757.350782] env[63088]: DEBUG nova.compute.manager [None req-1874938e-086f-4c3c-9049-55ea0246719c tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 757.379545] env[63088]: DEBUG nova.virt.hardware [None req-1874938e-086f-4c3c-9049-55ea0246719c tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 757.379545] env[63088]: DEBUG nova.virt.hardware [None req-1874938e-086f-4c3c-9049-55ea0246719c tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 757.379545] env[63088]: DEBUG nova.virt.hardware [None req-1874938e-086f-4c3c-9049-55ea0246719c tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 757.379786] env[63088]: DEBUG nova.virt.hardware [None req-1874938e-086f-4c3c-9049-55ea0246719c tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 757.379786] env[63088]: DEBUG nova.virt.hardware [None req-1874938e-086f-4c3c-9049-55ea0246719c tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 757.380303] env[63088]: DEBUG nova.virt.hardware [None req-1874938e-086f-4c3c-9049-55ea0246719c tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 757.380713] env[63088]: DEBUG nova.virt.hardware [None req-1874938e-086f-4c3c-9049-55ea0246719c tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 757.381052] env[63088]: DEBUG nova.virt.hardware [None req-1874938e-086f-4c3c-9049-55ea0246719c tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 757.384020] env[63088]: DEBUG nova.virt.hardware [None req-1874938e-086f-4c3c-9049-55ea0246719c tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 757.384020] env[63088]: DEBUG nova.virt.hardware [None req-1874938e-086f-4c3c-9049-55ea0246719c tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 757.384020] env[63088]: DEBUG nova.virt.hardware [None req-1874938e-086f-4c3c-9049-55ea0246719c tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 757.384020] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b5a9c22-1d87-4488-b4d1-0841fe65a8af {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.393571] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8d525e0-449e-4475-ad1e-b5bff1b36686 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.508867] env[63088]: DEBUG nova.network.neutron [None req-bce00a2e-7fb0-44a9-a66e-261b7c3d5384 tempest-ServerActionsV293TestJSON-1088890325 tempest-ServerActionsV293TestJSON-1088890325-project-member] [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 757.693316] env[63088]: DEBUG nova.compute.manager [req-e8269eaa-ac49-4d75-a807-7141fdf0bde1 req-fbdef190-d96e-41c2-8262-fb52c398a655 service nova] [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] Received event network-vif-deleted-685b39f8-6e13-43dc-9dfa-6be110e1ac12 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 757.748203] env[63088]: DEBUG nova.compute.manager [req-788e4343-4b0e-4bc0-9236-dc984d9709ac req-efe92568-2914-4c92-8c94-f76c2b76e042 service nova] [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] Received event network-changed-63967be2-f880-4637-9e38-8b244274427d {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 757.748203] env[63088]: DEBUG nova.compute.manager [req-788e4343-4b0e-4bc0-9236-dc984d9709ac req-efe92568-2914-4c92-8c94-f76c2b76e042 service nova] [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] Refreshing instance network info cache due to event network-changed-63967be2-f880-4637-9e38-8b244274427d. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 757.748203] env[63088]: DEBUG oslo_concurrency.lockutils [req-788e4343-4b0e-4bc0-9236-dc984d9709ac req-efe92568-2914-4c92-8c94-f76c2b76e042 service nova] Acquiring lock "refresh_cache-1e1732ef-a198-46e7-8f5e-3b9d8fe8c800" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 757.748535] env[63088]: DEBUG oslo_concurrency.lockutils [req-788e4343-4b0e-4bc0-9236-dc984d9709ac req-efe92568-2914-4c92-8c94-f76c2b76e042 service nova] Acquired lock "refresh_cache-1e1732ef-a198-46e7-8f5e-3b9d8fe8c800" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 757.748535] env[63088]: DEBUG nova.network.neutron [req-788e4343-4b0e-4bc0-9236-dc984d9709ac req-efe92568-2914-4c92-8c94-f76c2b76e042 service nova] [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] Refreshing network info cache for port 63967be2-f880-4637-9e38-8b244274427d {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 757.769181] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3edaa40f-e8ca-4f6f-a9f2-5744f490f65d tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.451s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 757.769756] env[63088]: DEBUG nova.compute.manager [None req-3edaa40f-e8ca-4f6f-a9f2-5744f490f65d tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 757.773772] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3a58af89-5e90-4bc8-9f8f-c02b80d14311 tempest-InstanceActionsTestJSON-1717540784 tempest-InstanceActionsTestJSON-1717540784-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.135s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 757.775622] env[63088]: INFO nova.compute.claims [None req-3a58af89-5e90-4bc8-9f8f-c02b80d14311 tempest-InstanceActionsTestJSON-1717540784 tempest-InstanceActionsTestJSON-1717540784-project-member] [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 757.948847] env[63088]: ERROR nova.compute.manager [None req-1874938e-086f-4c3c-9049-55ea0246719c tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 63967be2-f880-4637-9e38-8b244274427d, please check neutron logs for more information. [ 757.948847] env[63088]: ERROR nova.compute.manager Traceback (most recent call last): [ 757.948847] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 757.948847] env[63088]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 757.948847] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 757.948847] env[63088]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 757.948847] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 757.948847] env[63088]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 757.948847] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 757.948847] env[63088]: ERROR nova.compute.manager self.force_reraise() [ 757.948847] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 757.948847] env[63088]: ERROR nova.compute.manager raise self.value [ 757.948847] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 757.948847] env[63088]: ERROR nova.compute.manager updated_port = self._update_port( [ 757.948847] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 757.948847] env[63088]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 757.949672] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 757.949672] env[63088]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 757.949672] env[63088]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 63967be2-f880-4637-9e38-8b244274427d, please check neutron logs for more information. [ 757.949672] env[63088]: ERROR nova.compute.manager [ 757.949672] env[63088]: Traceback (most recent call last): [ 757.949672] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 757.949672] env[63088]: listener.cb(fileno) [ 757.949672] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 757.949672] env[63088]: result = function(*args, **kwargs) [ 757.949672] env[63088]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 757.949672] env[63088]: return func(*args, **kwargs) [ 757.949672] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 757.949672] env[63088]: raise e [ 757.949672] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 757.949672] env[63088]: nwinfo = self.network_api.allocate_for_instance( [ 757.949672] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 757.949672] env[63088]: created_port_ids = self._update_ports_for_instance( [ 757.949672] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 757.949672] env[63088]: with excutils.save_and_reraise_exception(): [ 757.949672] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 757.949672] env[63088]: self.force_reraise() [ 757.949672] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 757.949672] env[63088]: raise self.value [ 757.949672] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 757.949672] env[63088]: updated_port = self._update_port( [ 757.949672] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 757.949672] env[63088]: _ensure_no_port_binding_failure(port) [ 757.949672] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 757.949672] env[63088]: raise exception.PortBindingFailed(port_id=port['id']) [ 757.951689] env[63088]: nova.exception.PortBindingFailed: Binding failed for port 63967be2-f880-4637-9e38-8b244274427d, please check neutron logs for more information. [ 757.951689] env[63088]: Removing descriptor: 17 [ 757.951689] env[63088]: ERROR nova.compute.manager [None req-1874938e-086f-4c3c-9049-55ea0246719c tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 63967be2-f880-4637-9e38-8b244274427d, please check neutron logs for more information. [ 757.951689] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] Traceback (most recent call last): [ 757.951689] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 757.951689] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] yield resources [ 757.951689] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 757.951689] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] self.driver.spawn(context, instance, image_meta, [ 757.951689] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 757.951689] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] self._vmops.spawn(context, instance, image_meta, injected_files, [ 757.951689] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 757.951689] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] vm_ref = self.build_virtual_machine(instance, [ 757.952299] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 757.952299] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] vif_infos = vmwarevif.get_vif_info(self._session, [ 757.952299] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 757.952299] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] for vif in network_info: [ 757.952299] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 757.952299] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] return self._sync_wrapper(fn, *args, **kwargs) [ 757.952299] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 757.952299] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] self.wait() [ 757.952299] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 757.952299] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] self[:] = self._gt.wait() [ 757.952299] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 757.952299] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] return self._exit_event.wait() [ 757.952299] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 757.952862] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] result = hub.switch() [ 757.952862] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 757.952862] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] return self.greenlet.switch() [ 757.952862] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 757.952862] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] result = function(*args, **kwargs) [ 757.952862] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 757.952862] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] return func(*args, **kwargs) [ 757.952862] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 757.952862] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] raise e [ 757.952862] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 757.952862] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] nwinfo = self.network_api.allocate_for_instance( [ 757.952862] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 757.952862] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] created_port_ids = self._update_ports_for_instance( [ 757.953743] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 757.953743] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] with excutils.save_and_reraise_exception(): [ 757.953743] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 757.953743] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] self.force_reraise() [ 757.953743] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 757.953743] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] raise self.value [ 757.953743] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 757.953743] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] updated_port = self._update_port( [ 757.953743] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 757.953743] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] _ensure_no_port_binding_failure(port) [ 757.953743] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 757.953743] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] raise exception.PortBindingFailed(port_id=port['id']) [ 757.955110] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] nova.exception.PortBindingFailed: Binding failed for port 63967be2-f880-4637-9e38-8b244274427d, please check neutron logs for more information. [ 757.955110] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] [ 757.955110] env[63088]: INFO nova.compute.manager [None req-1874938e-086f-4c3c-9049-55ea0246719c tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] Terminating instance [ 757.957071] env[63088]: DEBUG oslo_concurrency.lockutils [None req-1874938e-086f-4c3c-9049-55ea0246719c tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] Acquiring lock "refresh_cache-1e1732ef-a198-46e7-8f5e-3b9d8fe8c800" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 758.012023] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bce00a2e-7fb0-44a9-a66e-261b7c3d5384 tempest-ServerActionsV293TestJSON-1088890325 tempest-ServerActionsV293TestJSON-1088890325-project-member] Releasing lock "refresh_cache-b82c89ad-f7d6-4c04-a48e-acdab29ea5e7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 758.013401] env[63088]: DEBUG nova.compute.manager [None req-bce00a2e-7fb0-44a9-a66e-261b7c3d5384 tempest-ServerActionsV293TestJSON-1088890325 tempest-ServerActionsV293TestJSON-1088890325-project-member] [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 758.013724] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-277408b8-aa96-45fe-b8b4-83c42f262b4f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.024157] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18a0286b-ac7a-4dd2-8c11-d5e71cbf8534 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.048328] env[63088]: WARNING nova.virt.vmwareapi.driver [None req-bce00a2e-7fb0-44a9-a66e-261b7c3d5384 tempest-ServerActionsV293TestJSON-1088890325 tempest-ServerActionsV293TestJSON-1088890325-project-member] [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance b82c89ad-f7d6-4c04-a48e-acdab29ea5e7 could not be found. [ 758.048328] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-bce00a2e-7fb0-44a9-a66e-261b7c3d5384 tempest-ServerActionsV293TestJSON-1088890325 tempest-ServerActionsV293TestJSON-1088890325-project-member] [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 758.048479] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f8628835-8c08-4451-b15f-39781c63a165 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.056778] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce88db22-2b7b-4406-9958-54dd970973dc {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.088780] env[63088]: WARNING nova.virt.vmwareapi.vmops [None req-bce00a2e-7fb0-44a9-a66e-261b7c3d5384 tempest-ServerActionsV293TestJSON-1088890325 tempest-ServerActionsV293TestJSON-1088890325-project-member] [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b82c89ad-f7d6-4c04-a48e-acdab29ea5e7 could not be found. [ 758.089081] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-bce00a2e-7fb0-44a9-a66e-261b7c3d5384 tempest-ServerActionsV293TestJSON-1088890325 tempest-ServerActionsV293TestJSON-1088890325-project-member] [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 758.089286] env[63088]: INFO nova.compute.manager [None req-bce00a2e-7fb0-44a9-a66e-261b7c3d5384 tempest-ServerActionsV293TestJSON-1088890325 tempest-ServerActionsV293TestJSON-1088890325-project-member] [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] Took 0.08 seconds to destroy the instance on the hypervisor. [ 758.089576] env[63088]: DEBUG oslo.service.loopingcall [None req-bce00a2e-7fb0-44a9-a66e-261b7c3d5384 tempest-ServerActionsV293TestJSON-1088890325 tempest-ServerActionsV293TestJSON-1088890325-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 758.089818] env[63088]: DEBUG nova.compute.manager [-] [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 758.089912] env[63088]: DEBUG nova.network.neutron [-] [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 758.107304] env[63088]: DEBUG nova.network.neutron [-] [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 758.277118] env[63088]: DEBUG nova.network.neutron [req-788e4343-4b0e-4bc0-9236-dc984d9709ac req-efe92568-2914-4c92-8c94-f76c2b76e042 service nova] [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 758.278744] env[63088]: DEBUG nova.compute.utils [None req-3edaa40f-e8ca-4f6f-a9f2-5744f490f65d tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 758.280754] env[63088]: DEBUG nova.compute.manager [None req-3edaa40f-e8ca-4f6f-a9f2-5744f490f65d tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 758.281116] env[63088]: DEBUG nova.network.neutron [None req-3edaa40f-e8ca-4f6f-a9f2-5744f490f65d tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 758.366057] env[63088]: DEBUG nova.network.neutron [req-788e4343-4b0e-4bc0-9236-dc984d9709ac req-efe92568-2914-4c92-8c94-f76c2b76e042 service nova] [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 758.390192] env[63088]: DEBUG nova.policy [None req-3edaa40f-e8ca-4f6f-a9f2-5744f490f65d tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c54bbb9377894ca4bfd98d74dab12c6f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '18e424c0af324c33ad0b94a268c4651d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 758.611772] env[63088]: DEBUG nova.network.neutron [-] [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 758.790724] env[63088]: DEBUG nova.compute.manager [None req-3edaa40f-e8ca-4f6f-a9f2-5744f490f65d tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 758.870475] env[63088]: DEBUG oslo_concurrency.lockutils [req-788e4343-4b0e-4bc0-9236-dc984d9709ac req-efe92568-2914-4c92-8c94-f76c2b76e042 service nova] Releasing lock "refresh_cache-1e1732ef-a198-46e7-8f5e-3b9d8fe8c800" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 758.873384] env[63088]: DEBUG oslo_concurrency.lockutils [None req-1874938e-086f-4c3c-9049-55ea0246719c tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] Acquired lock "refresh_cache-1e1732ef-a198-46e7-8f5e-3b9d8fe8c800" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 758.874251] env[63088]: DEBUG nova.network.neutron [None req-1874938e-086f-4c3c-9049-55ea0246719c tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 758.949275] env[63088]: DEBUG nova.network.neutron [None req-3edaa40f-e8ca-4f6f-a9f2-5744f490f65d tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] Successfully created port: 70b3718f-5593-4ace-be43-0e092ff9bfe7 {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 759.114513] env[63088]: INFO nova.compute.manager [-] [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] Took 1.02 seconds to deallocate network for instance. [ 759.120731] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4632e797-725b-4c9d-8e76-a9cbee72b45f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.129266] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1362af7d-d9f1-4f06-b973-b3a3519b4385 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.163377] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d311df09-74aa-4367-9038-49decfc1db3d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.171380] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-622bb7a4-83c2-446a-ae69-5342ea79bdab {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.186549] env[63088]: DEBUG nova.compute.provider_tree [None req-3a58af89-5e90-4bc8-9f8f-c02b80d14311 tempest-InstanceActionsTestJSON-1717540784 tempest-InstanceActionsTestJSON-1717540784-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 759.393844] env[63088]: DEBUG nova.network.neutron [None req-1874938e-086f-4c3c-9049-55ea0246719c tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 759.539511] env[63088]: DEBUG nova.network.neutron [None req-1874938e-086f-4c3c-9049-55ea0246719c tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 759.690522] env[63088]: DEBUG nova.scheduler.client.report [None req-3a58af89-5e90-4bc8-9f8f-c02b80d14311 tempest-InstanceActionsTestJSON-1717540784 tempest-InstanceActionsTestJSON-1717540784-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 759.694229] env[63088]: INFO nova.compute.manager [None req-bce00a2e-7fb0-44a9-a66e-261b7c3d5384 tempest-ServerActionsV293TestJSON-1088890325 tempest-ServerActionsV293TestJSON-1088890325-project-member] [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] Took 0.58 seconds to detach 1 volumes for instance. [ 759.696283] env[63088]: DEBUG nova.compute.claims [None req-bce00a2e-7fb0-44a9-a66e-261b7c3d5384 tempest-ServerActionsV293TestJSON-1088890325 tempest-ServerActionsV293TestJSON-1088890325-project-member] [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] Aborting claim: {{(pid=63088) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 759.696283] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bce00a2e-7fb0-44a9-a66e-261b7c3d5384 tempest-ServerActionsV293TestJSON-1088890325 tempest-ServerActionsV293TestJSON-1088890325-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 759.774576] env[63088]: DEBUG nova.compute.manager [req-aa81c16c-8aa5-471a-b4d8-3302663ac691 req-db673e37-1424-4162-8bca-d8a8e661411a service nova] [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] Received event network-vif-deleted-63967be2-f880-4637-9e38-8b244274427d {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 759.803629] env[63088]: DEBUG nova.compute.manager [None req-3edaa40f-e8ca-4f6f-a9f2-5744f490f65d tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 759.833289] env[63088]: DEBUG nova.virt.hardware [None req-3edaa40f-e8ca-4f6f-a9f2-5744f490f65d tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 759.833519] env[63088]: DEBUG nova.virt.hardware [None req-3edaa40f-e8ca-4f6f-a9f2-5744f490f65d tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 759.833680] env[63088]: DEBUG nova.virt.hardware [None req-3edaa40f-e8ca-4f6f-a9f2-5744f490f65d tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 759.833913] env[63088]: DEBUG nova.virt.hardware [None req-3edaa40f-e8ca-4f6f-a9f2-5744f490f65d tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 759.834010] env[63088]: DEBUG nova.virt.hardware [None req-3edaa40f-e8ca-4f6f-a9f2-5744f490f65d tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 759.838020] env[63088]: DEBUG nova.virt.hardware [None req-3edaa40f-e8ca-4f6f-a9f2-5744f490f65d tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 759.838020] env[63088]: DEBUG nova.virt.hardware [None req-3edaa40f-e8ca-4f6f-a9f2-5744f490f65d tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 759.838020] env[63088]: DEBUG nova.virt.hardware [None req-3edaa40f-e8ca-4f6f-a9f2-5744f490f65d tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 759.838020] env[63088]: DEBUG nova.virt.hardware [None req-3edaa40f-e8ca-4f6f-a9f2-5744f490f65d tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 759.838020] env[63088]: DEBUG nova.virt.hardware [None req-3edaa40f-e8ca-4f6f-a9f2-5744f490f65d tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 759.838393] env[63088]: DEBUG nova.virt.hardware [None req-3edaa40f-e8ca-4f6f-a9f2-5744f490f65d tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 759.838393] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b63645d1-3a87-4be2-8e89-8d8dc1cb4e9d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.849616] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5992ce92-5847-415d-99e3-74b56be8c399 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.000196] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 760.000459] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 760.042138] env[63088]: DEBUG oslo_concurrency.lockutils [None req-1874938e-086f-4c3c-9049-55ea0246719c tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] Releasing lock "refresh_cache-1e1732ef-a198-46e7-8f5e-3b9d8fe8c800" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 760.042445] env[63088]: DEBUG nova.compute.manager [None req-1874938e-086f-4c3c-9049-55ea0246719c tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 760.042651] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-1874938e-086f-4c3c-9049-55ea0246719c tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 760.042934] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9d2b857e-e24b-4c2f-920c-0552cf5f107d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.053524] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-922be2b0-3582-4420-aea6-108ef8ae5bfd {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.077142] env[63088]: WARNING nova.virt.vmwareapi.vmops [None req-1874938e-086f-4c3c-9049-55ea0246719c tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800 could not be found. [ 760.077339] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-1874938e-086f-4c3c-9049-55ea0246719c tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 760.077530] env[63088]: INFO nova.compute.manager [None req-1874938e-086f-4c3c-9049-55ea0246719c tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] Took 0.03 seconds to destroy the instance on the hypervisor. [ 760.078424] env[63088]: DEBUG oslo.service.loopingcall [None req-1874938e-086f-4c3c-9049-55ea0246719c tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 760.078966] env[63088]: DEBUG nova.compute.manager [-] [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 760.079098] env[63088]: DEBUG nova.network.neutron [-] [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 760.097643] env[63088]: DEBUG nova.network.neutron [-] [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 760.197143] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3a58af89-5e90-4bc8-9f8f-c02b80d14311 tempest-InstanceActionsTestJSON-1717540784 tempest-InstanceActionsTestJSON-1717540784-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.423s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 760.197562] env[63088]: DEBUG nova.compute.manager [None req-3a58af89-5e90-4bc8-9f8f-c02b80d14311 tempest-InstanceActionsTestJSON-1717540784 tempest-InstanceActionsTestJSON-1717540784-project-member] [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 760.210488] env[63088]: DEBUG oslo_concurrency.lockutils [None req-87ab5d07-9ebb-429d-8640-0197960e7d1a tempest-ServersNegativeTestMultiTenantJSON-2054426678 tempest-ServersNegativeTestMultiTenantJSON-2054426678-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.713s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 760.212963] env[63088]: INFO nova.compute.claims [None req-87ab5d07-9ebb-429d-8640-0197960e7d1a tempest-ServersNegativeTestMultiTenantJSON-2054426678 tempest-ServersNegativeTestMultiTenantJSON-2054426678-project-member] [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 760.385597] env[63088]: DEBUG nova.compute.manager [req-ec7e8109-6a04-40ff-984a-834eb35cb3b3 req-9adb3002-30cb-4454-b2a2-2a4d10f7b136 service nova] [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] Received event network-changed-70b3718f-5593-4ace-be43-0e092ff9bfe7 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 760.385802] env[63088]: DEBUG nova.compute.manager [req-ec7e8109-6a04-40ff-984a-834eb35cb3b3 req-9adb3002-30cb-4454-b2a2-2a4d10f7b136 service nova] [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] Refreshing instance network info cache due to event network-changed-70b3718f-5593-4ace-be43-0e092ff9bfe7. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 760.386029] env[63088]: DEBUG oslo_concurrency.lockutils [req-ec7e8109-6a04-40ff-984a-834eb35cb3b3 req-9adb3002-30cb-4454-b2a2-2a4d10f7b136 service nova] Acquiring lock "refresh_cache-17725a7b-ccd3-477e-a7a6-b96af45d6d1f" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 760.386156] env[63088]: DEBUG oslo_concurrency.lockutils [req-ec7e8109-6a04-40ff-984a-834eb35cb3b3 req-9adb3002-30cb-4454-b2a2-2a4d10f7b136 service nova] Acquired lock "refresh_cache-17725a7b-ccd3-477e-a7a6-b96af45d6d1f" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 760.386317] env[63088]: DEBUG nova.network.neutron [req-ec7e8109-6a04-40ff-984a-834eb35cb3b3 req-9adb3002-30cb-4454-b2a2-2a4d10f7b136 service nova] [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] Refreshing network info cache for port 70b3718f-5593-4ace-be43-0e092ff9bfe7 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 760.510160] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 760.510455] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Starting heal instance info cache {{(pid=63088) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 760.510572] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Rebuilding the list of instances to heal {{(pid=63088) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 760.601937] env[63088]: DEBUG nova.network.neutron [-] [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 760.672364] env[63088]: ERROR nova.compute.manager [None req-3edaa40f-e8ca-4f6f-a9f2-5744f490f65d tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 70b3718f-5593-4ace-be43-0e092ff9bfe7, please check neutron logs for more information. [ 760.672364] env[63088]: ERROR nova.compute.manager Traceback (most recent call last): [ 760.672364] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 760.672364] env[63088]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 760.672364] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 760.672364] env[63088]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 760.672364] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 760.672364] env[63088]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 760.672364] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 760.672364] env[63088]: ERROR nova.compute.manager self.force_reraise() [ 760.672364] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 760.672364] env[63088]: ERROR nova.compute.manager raise self.value [ 760.672364] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 760.672364] env[63088]: ERROR nova.compute.manager updated_port = self._update_port( [ 760.672364] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 760.672364] env[63088]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 760.672828] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 760.672828] env[63088]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 760.672828] env[63088]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 70b3718f-5593-4ace-be43-0e092ff9bfe7, please check neutron logs for more information. [ 760.672828] env[63088]: ERROR nova.compute.manager [ 760.672828] env[63088]: Traceback (most recent call last): [ 760.672828] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 760.672828] env[63088]: listener.cb(fileno) [ 760.672828] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 760.672828] env[63088]: result = function(*args, **kwargs) [ 760.672828] env[63088]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 760.672828] env[63088]: return func(*args, **kwargs) [ 760.672828] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 760.672828] env[63088]: raise e [ 760.672828] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 760.672828] env[63088]: nwinfo = self.network_api.allocate_for_instance( [ 760.672828] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 760.672828] env[63088]: created_port_ids = self._update_ports_for_instance( [ 760.672828] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 760.672828] env[63088]: with excutils.save_and_reraise_exception(): [ 760.672828] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 760.672828] env[63088]: self.force_reraise() [ 760.672828] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 760.672828] env[63088]: raise self.value [ 760.672828] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 760.672828] env[63088]: updated_port = self._update_port( [ 760.672828] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 760.672828] env[63088]: _ensure_no_port_binding_failure(port) [ 760.672828] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 760.672828] env[63088]: raise exception.PortBindingFailed(port_id=port['id']) [ 760.673599] env[63088]: nova.exception.PortBindingFailed: Binding failed for port 70b3718f-5593-4ace-be43-0e092ff9bfe7, please check neutron logs for more information. [ 760.673599] env[63088]: Removing descriptor: 15 [ 760.673599] env[63088]: ERROR nova.compute.manager [None req-3edaa40f-e8ca-4f6f-a9f2-5744f490f65d tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 70b3718f-5593-4ace-be43-0e092ff9bfe7, please check neutron logs for more information. [ 760.673599] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] Traceback (most recent call last): [ 760.673599] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 760.673599] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] yield resources [ 760.673599] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 760.673599] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] self.driver.spawn(context, instance, image_meta, [ 760.673599] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 760.673599] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 760.673599] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 760.673599] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] vm_ref = self.build_virtual_machine(instance, [ 760.673967] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 760.673967] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] vif_infos = vmwarevif.get_vif_info(self._session, [ 760.673967] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 760.673967] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] for vif in network_info: [ 760.673967] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 760.673967] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] return self._sync_wrapper(fn, *args, **kwargs) [ 760.673967] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 760.673967] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] self.wait() [ 760.673967] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 760.673967] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] self[:] = self._gt.wait() [ 760.673967] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 760.673967] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] return self._exit_event.wait() [ 760.673967] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 760.674362] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] result = hub.switch() [ 760.674362] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 760.674362] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] return self.greenlet.switch() [ 760.674362] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 760.674362] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] result = function(*args, **kwargs) [ 760.674362] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 760.674362] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] return func(*args, **kwargs) [ 760.674362] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 760.674362] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] raise e [ 760.674362] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 760.674362] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] nwinfo = self.network_api.allocate_for_instance( [ 760.674362] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 760.674362] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] created_port_ids = self._update_ports_for_instance( [ 760.674716] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 760.674716] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] with excutils.save_and_reraise_exception(): [ 760.674716] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 760.674716] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] self.force_reraise() [ 760.674716] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 760.674716] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] raise self.value [ 760.674716] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 760.674716] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] updated_port = self._update_port( [ 760.674716] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 760.674716] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] _ensure_no_port_binding_failure(port) [ 760.674716] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 760.674716] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] raise exception.PortBindingFailed(port_id=port['id']) [ 760.675107] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] nova.exception.PortBindingFailed: Binding failed for port 70b3718f-5593-4ace-be43-0e092ff9bfe7, please check neutron logs for more information. [ 760.675107] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] [ 760.675107] env[63088]: INFO nova.compute.manager [None req-3edaa40f-e8ca-4f6f-a9f2-5744f490f65d tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] Terminating instance [ 760.676288] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3edaa40f-e8ca-4f6f-a9f2-5744f490f65d tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] Acquiring lock "refresh_cache-17725a7b-ccd3-477e-a7a6-b96af45d6d1f" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 760.715568] env[63088]: DEBUG nova.compute.utils [None req-3a58af89-5e90-4bc8-9f8f-c02b80d14311 tempest-InstanceActionsTestJSON-1717540784 tempest-InstanceActionsTestJSON-1717540784-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 760.716924] env[63088]: DEBUG nova.compute.manager [None req-3a58af89-5e90-4bc8-9f8f-c02b80d14311 tempest-InstanceActionsTestJSON-1717540784 tempest-InstanceActionsTestJSON-1717540784-project-member] [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 760.717074] env[63088]: DEBUG nova.network.neutron [None req-3a58af89-5e90-4bc8-9f8f-c02b80d14311 tempest-InstanceActionsTestJSON-1717540784 tempest-InstanceActionsTestJSON-1717540784-project-member] [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 760.777271] env[63088]: DEBUG nova.policy [None req-3a58af89-5e90-4bc8-9f8f-c02b80d14311 tempest-InstanceActionsTestJSON-1717540784 tempest-InstanceActionsTestJSON-1717540784-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '27b443b30b8a498b9d25a2eb2d729246', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5a31e233404947cfa3926fa0aef006bb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 760.909777] env[63088]: DEBUG nova.network.neutron [req-ec7e8109-6a04-40ff-984a-834eb35cb3b3 req-9adb3002-30cb-4454-b2a2-2a4d10f7b136 service nova] [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 761.014900] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] Skipping network cache update for instance because it is Building. {{(pid=63088) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 761.015372] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] Skipping network cache update for instance because it is Building. {{(pid=63088) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 761.015372] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] Skipping network cache update for instance because it is Building. {{(pid=63088) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 761.015372] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] Skipping network cache update for instance because it is Building. {{(pid=63088) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 761.015558] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] Skipping network cache update for instance because it is Building. {{(pid=63088) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 761.015608] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] Skipping network cache update for instance because it is Building. {{(pid=63088) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 761.015737] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Didn't find any instances for network info cache update. {{(pid=63088) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10012}} [ 761.015911] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 761.016301] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 761.016626] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 761.016758] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 761.016882] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 761.017220] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 761.017388] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63088) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 761.017664] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager.update_available_resource {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 761.083243] env[63088]: DEBUG nova.network.neutron [req-ec7e8109-6a04-40ff-984a-834eb35cb3b3 req-9adb3002-30cb-4454-b2a2-2a4d10f7b136 service nova] [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 761.103791] env[63088]: INFO nova.compute.manager [-] [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] Took 1.02 seconds to deallocate network for instance. [ 761.106067] env[63088]: DEBUG nova.compute.claims [None req-1874938e-086f-4c3c-9049-55ea0246719c tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] Aborting claim: {{(pid=63088) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 761.106274] env[63088]: DEBUG oslo_concurrency.lockutils [None req-1874938e-086f-4c3c-9049-55ea0246719c tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 761.161866] env[63088]: DEBUG nova.network.neutron [None req-3a58af89-5e90-4bc8-9f8f-c02b80d14311 tempest-InstanceActionsTestJSON-1717540784 tempest-InstanceActionsTestJSON-1717540784-project-member] [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] Successfully created port: 6904053f-859c-4b6f-80fc-0278d414bf42 {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 761.222968] env[63088]: DEBUG nova.compute.manager [None req-3a58af89-5e90-4bc8-9f8f-c02b80d14311 tempest-InstanceActionsTestJSON-1717540784 tempest-InstanceActionsTestJSON-1717540784-project-member] [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 761.520559] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 761.525070] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea36b29a-85e3-41b4-a78c-809b124a29ad {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.533205] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-743a0c9f-c6e2-43c4-8599-418bda0822d5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.564511] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b7b271d-2387-40ec-be03-da2d6f9a80a5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.571741] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a4e64b2-cd2e-443f-af14-8f7526124734 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.585225] env[63088]: DEBUG nova.compute.provider_tree [None req-87ab5d07-9ebb-429d-8640-0197960e7d1a tempest-ServersNegativeTestMultiTenantJSON-2054426678 tempest-ServersNegativeTestMultiTenantJSON-2054426678-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 761.591104] env[63088]: DEBUG oslo_concurrency.lockutils [req-ec7e8109-6a04-40ff-984a-834eb35cb3b3 req-9adb3002-30cb-4454-b2a2-2a4d10f7b136 service nova] Releasing lock "refresh_cache-17725a7b-ccd3-477e-a7a6-b96af45d6d1f" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 761.591472] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3edaa40f-e8ca-4f6f-a9f2-5744f490f65d tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] Acquired lock "refresh_cache-17725a7b-ccd3-477e-a7a6-b96af45d6d1f" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 761.591646] env[63088]: DEBUG nova.network.neutron [None req-3edaa40f-e8ca-4f6f-a9f2-5744f490f65d tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 762.091198] env[63088]: DEBUG nova.scheduler.client.report [None req-87ab5d07-9ebb-429d-8640-0197960e7d1a tempest-ServersNegativeTestMultiTenantJSON-2054426678 tempest-ServersNegativeTestMultiTenantJSON-2054426678-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 762.120888] env[63088]: DEBUG nova.network.neutron [None req-3edaa40f-e8ca-4f6f-a9f2-5744f490f65d tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 762.222087] env[63088]: DEBUG nova.network.neutron [None req-3edaa40f-e8ca-4f6f-a9f2-5744f490f65d tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 762.239704] env[63088]: DEBUG nova.compute.manager [None req-3a58af89-5e90-4bc8-9f8f-c02b80d14311 tempest-InstanceActionsTestJSON-1717540784 tempest-InstanceActionsTestJSON-1717540784-project-member] [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 762.275837] env[63088]: DEBUG nova.virt.hardware [None req-3a58af89-5e90-4bc8-9f8f-c02b80d14311 tempest-InstanceActionsTestJSON-1717540784 tempest-InstanceActionsTestJSON-1717540784-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 762.276103] env[63088]: DEBUG nova.virt.hardware [None req-3a58af89-5e90-4bc8-9f8f-c02b80d14311 tempest-InstanceActionsTestJSON-1717540784 tempest-InstanceActionsTestJSON-1717540784-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 762.276275] env[63088]: DEBUG nova.virt.hardware [None req-3a58af89-5e90-4bc8-9f8f-c02b80d14311 tempest-InstanceActionsTestJSON-1717540784 tempest-InstanceActionsTestJSON-1717540784-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 762.276462] env[63088]: DEBUG nova.virt.hardware [None req-3a58af89-5e90-4bc8-9f8f-c02b80d14311 tempest-InstanceActionsTestJSON-1717540784 tempest-InstanceActionsTestJSON-1717540784-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 762.276607] env[63088]: DEBUG nova.virt.hardware [None req-3a58af89-5e90-4bc8-9f8f-c02b80d14311 tempest-InstanceActionsTestJSON-1717540784 tempest-InstanceActionsTestJSON-1717540784-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 762.276750] env[63088]: DEBUG nova.virt.hardware [None req-3a58af89-5e90-4bc8-9f8f-c02b80d14311 tempest-InstanceActionsTestJSON-1717540784 tempest-InstanceActionsTestJSON-1717540784-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 762.276949] env[63088]: DEBUG nova.virt.hardware [None req-3a58af89-5e90-4bc8-9f8f-c02b80d14311 tempest-InstanceActionsTestJSON-1717540784 tempest-InstanceActionsTestJSON-1717540784-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 762.277218] env[63088]: DEBUG nova.virt.hardware [None req-3a58af89-5e90-4bc8-9f8f-c02b80d14311 tempest-InstanceActionsTestJSON-1717540784 tempest-InstanceActionsTestJSON-1717540784-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 762.277466] env[63088]: DEBUG nova.virt.hardware [None req-3a58af89-5e90-4bc8-9f8f-c02b80d14311 tempest-InstanceActionsTestJSON-1717540784 tempest-InstanceActionsTestJSON-1717540784-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 762.277598] env[63088]: DEBUG nova.virt.hardware [None req-3a58af89-5e90-4bc8-9f8f-c02b80d14311 tempest-InstanceActionsTestJSON-1717540784 tempest-InstanceActionsTestJSON-1717540784-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 762.277770] env[63088]: DEBUG nova.virt.hardware [None req-3a58af89-5e90-4bc8-9f8f-c02b80d14311 tempest-InstanceActionsTestJSON-1717540784 tempest-InstanceActionsTestJSON-1717540784-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 762.278637] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07973006-3b27-4690-8e9b-ba4ad9ff99e0 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.292192] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa955c14-d394-4b73-865e-6e4b37fa21dd {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.551148] env[63088]: DEBUG nova.compute.manager [req-a7e8d54b-df8a-4d7b-8c72-f9a6a0e51e9a req-78654444-7ca7-4eec-af2c-b787c929cbfd service nova] [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] Received event network-vif-deleted-70b3718f-5593-4ace-be43-0e092ff9bfe7 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 762.597870] env[63088]: DEBUG oslo_concurrency.lockutils [None req-87ab5d07-9ebb-429d-8640-0197960e7d1a tempest-ServersNegativeTestMultiTenantJSON-2054426678 tempest-ServersNegativeTestMultiTenantJSON-2054426678-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.387s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 762.598444] env[63088]: DEBUG nova.compute.manager [None req-87ab5d07-9ebb-429d-8640-0197960e7d1a tempest-ServersNegativeTestMultiTenantJSON-2054426678 tempest-ServersNegativeTestMultiTenantJSON-2054426678-project-member] [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 762.602035] env[63088]: DEBUG oslo_concurrency.lockutils [None req-aa39901a-ab16-4a54-a2f6-147052a0ed90 tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.270s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 762.606378] env[63088]: INFO nova.compute.claims [None req-aa39901a-ab16-4a54-a2f6-147052a0ed90 tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] [instance: 43ded929-6890-4296-ae90-230f7c8b2595] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 762.724899] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3edaa40f-e8ca-4f6f-a9f2-5744f490f65d tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] Releasing lock "refresh_cache-17725a7b-ccd3-477e-a7a6-b96af45d6d1f" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 762.725363] env[63088]: DEBUG nova.compute.manager [None req-3edaa40f-e8ca-4f6f-a9f2-5744f490f65d tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 762.725590] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-3edaa40f-e8ca-4f6f-a9f2-5744f490f65d tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 762.726079] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-24ea765c-7b5e-4e54-a154-0dba8292dd76 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.735728] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4b63ca5-3b01-411c-842d-0dfaede62163 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.759638] env[63088]: WARNING nova.virt.vmwareapi.vmops [None req-3edaa40f-e8ca-4f6f-a9f2-5744f490f65d tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 17725a7b-ccd3-477e-a7a6-b96af45d6d1f could not be found. [ 762.759863] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-3edaa40f-e8ca-4f6f-a9f2-5744f490f65d tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 762.760059] env[63088]: INFO nova.compute.manager [None req-3edaa40f-e8ca-4f6f-a9f2-5744f490f65d tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] Took 0.03 seconds to destroy the instance on the hypervisor. [ 762.760335] env[63088]: DEBUG oslo.service.loopingcall [None req-3edaa40f-e8ca-4f6f-a9f2-5744f490f65d tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 762.760559] env[63088]: DEBUG nova.compute.manager [-] [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 762.760654] env[63088]: DEBUG nova.network.neutron [-] [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 762.777436] env[63088]: DEBUG nova.network.neutron [-] [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 762.869862] env[63088]: ERROR nova.compute.manager [None req-3a58af89-5e90-4bc8-9f8f-c02b80d14311 tempest-InstanceActionsTestJSON-1717540784 tempest-InstanceActionsTestJSON-1717540784-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6904053f-859c-4b6f-80fc-0278d414bf42, please check neutron logs for more information. [ 762.869862] env[63088]: ERROR nova.compute.manager Traceback (most recent call last): [ 762.869862] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 762.869862] env[63088]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 762.869862] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 762.869862] env[63088]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 762.869862] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 762.869862] env[63088]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 762.869862] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 762.869862] env[63088]: ERROR nova.compute.manager self.force_reraise() [ 762.869862] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 762.869862] env[63088]: ERROR nova.compute.manager raise self.value [ 762.869862] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 762.869862] env[63088]: ERROR nova.compute.manager updated_port = self._update_port( [ 762.869862] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 762.869862] env[63088]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 762.870432] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 762.870432] env[63088]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 762.870432] env[63088]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6904053f-859c-4b6f-80fc-0278d414bf42, please check neutron logs for more information. [ 762.870432] env[63088]: ERROR nova.compute.manager [ 762.870432] env[63088]: Traceback (most recent call last): [ 762.870432] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 762.870432] env[63088]: listener.cb(fileno) [ 762.870432] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 762.870432] env[63088]: result = function(*args, **kwargs) [ 762.870432] env[63088]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 762.870432] env[63088]: return func(*args, **kwargs) [ 762.870432] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 762.870432] env[63088]: raise e [ 762.870432] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 762.870432] env[63088]: nwinfo = self.network_api.allocate_for_instance( [ 762.870432] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 762.870432] env[63088]: created_port_ids = self._update_ports_for_instance( [ 762.870432] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 762.870432] env[63088]: with excutils.save_and_reraise_exception(): [ 762.870432] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 762.870432] env[63088]: self.force_reraise() [ 762.870432] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 762.870432] env[63088]: raise self.value [ 762.870432] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 762.870432] env[63088]: updated_port = self._update_port( [ 762.870432] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 762.870432] env[63088]: _ensure_no_port_binding_failure(port) [ 762.870432] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 762.870432] env[63088]: raise exception.PortBindingFailed(port_id=port['id']) [ 762.871274] env[63088]: nova.exception.PortBindingFailed: Binding failed for port 6904053f-859c-4b6f-80fc-0278d414bf42, please check neutron logs for more information. [ 762.871274] env[63088]: Removing descriptor: 15 [ 762.871274] env[63088]: ERROR nova.compute.manager [None req-3a58af89-5e90-4bc8-9f8f-c02b80d14311 tempest-InstanceActionsTestJSON-1717540784 tempest-InstanceActionsTestJSON-1717540784-project-member] [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6904053f-859c-4b6f-80fc-0278d414bf42, please check neutron logs for more information. [ 762.871274] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] Traceback (most recent call last): [ 762.871274] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 762.871274] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] yield resources [ 762.871274] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 762.871274] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] self.driver.spawn(context, instance, image_meta, [ 762.871274] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 762.871274] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] self._vmops.spawn(context, instance, image_meta, injected_files, [ 762.871274] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 762.871274] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] vm_ref = self.build_virtual_machine(instance, [ 762.871668] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 762.871668] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] vif_infos = vmwarevif.get_vif_info(self._session, [ 762.871668] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 762.871668] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] for vif in network_info: [ 762.871668] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 762.871668] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] return self._sync_wrapper(fn, *args, **kwargs) [ 762.871668] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 762.871668] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] self.wait() [ 762.871668] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 762.871668] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] self[:] = self._gt.wait() [ 762.871668] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 762.871668] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] return self._exit_event.wait() [ 762.871668] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 762.872072] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] result = hub.switch() [ 762.872072] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 762.872072] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] return self.greenlet.switch() [ 762.872072] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 762.872072] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] result = function(*args, **kwargs) [ 762.872072] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 762.872072] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] return func(*args, **kwargs) [ 762.872072] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 762.872072] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] raise e [ 762.872072] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 762.872072] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] nwinfo = self.network_api.allocate_for_instance( [ 762.872072] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 762.872072] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] created_port_ids = self._update_ports_for_instance( [ 762.872510] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 762.872510] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] with excutils.save_and_reraise_exception(): [ 762.872510] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 762.872510] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] self.force_reraise() [ 762.872510] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 762.872510] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] raise self.value [ 762.872510] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 762.872510] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] updated_port = self._update_port( [ 762.872510] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 762.872510] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] _ensure_no_port_binding_failure(port) [ 762.872510] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 762.872510] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] raise exception.PortBindingFailed(port_id=port['id']) [ 762.872897] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] nova.exception.PortBindingFailed: Binding failed for port 6904053f-859c-4b6f-80fc-0278d414bf42, please check neutron logs for more information. [ 762.872897] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] [ 762.872897] env[63088]: INFO nova.compute.manager [None req-3a58af89-5e90-4bc8-9f8f-c02b80d14311 tempest-InstanceActionsTestJSON-1717540784 tempest-InstanceActionsTestJSON-1717540784-project-member] [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] Terminating instance [ 762.875990] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3a58af89-5e90-4bc8-9f8f-c02b80d14311 tempest-InstanceActionsTestJSON-1717540784 tempest-InstanceActionsTestJSON-1717540784-project-member] Acquiring lock "refresh_cache-7b5cc7b5-98b9-4344-b714-270dca798799" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 762.876170] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3a58af89-5e90-4bc8-9f8f-c02b80d14311 tempest-InstanceActionsTestJSON-1717540784 tempest-InstanceActionsTestJSON-1717540784-project-member] Acquired lock "refresh_cache-7b5cc7b5-98b9-4344-b714-270dca798799" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 762.876340] env[63088]: DEBUG nova.network.neutron [None req-3a58af89-5e90-4bc8-9f8f-c02b80d14311 tempest-InstanceActionsTestJSON-1717540784 tempest-InstanceActionsTestJSON-1717540784-project-member] [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 763.112136] env[63088]: DEBUG nova.compute.utils [None req-87ab5d07-9ebb-429d-8640-0197960e7d1a tempest-ServersNegativeTestMultiTenantJSON-2054426678 tempest-ServersNegativeTestMultiTenantJSON-2054426678-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 763.114032] env[63088]: DEBUG nova.compute.manager [None req-87ab5d07-9ebb-429d-8640-0197960e7d1a tempest-ServersNegativeTestMultiTenantJSON-2054426678 tempest-ServersNegativeTestMultiTenantJSON-2054426678-project-member] [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 763.114169] env[63088]: DEBUG nova.network.neutron [None req-87ab5d07-9ebb-429d-8640-0197960e7d1a tempest-ServersNegativeTestMultiTenantJSON-2054426678 tempest-ServersNegativeTestMultiTenantJSON-2054426678-project-member] [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 763.173528] env[63088]: DEBUG nova.policy [None req-87ab5d07-9ebb-429d-8640-0197960e7d1a tempest-ServersNegativeTestMultiTenantJSON-2054426678 tempest-ServersNegativeTestMultiTenantJSON-2054426678-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bc1666929c1a45d688c800dee8ed57ce', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'eddfc8f3f77c406eb4d02dc9700441f1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 763.280702] env[63088]: DEBUG nova.network.neutron [-] [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 763.399341] env[63088]: DEBUG nova.network.neutron [None req-3a58af89-5e90-4bc8-9f8f-c02b80d14311 tempest-InstanceActionsTestJSON-1717540784 tempest-InstanceActionsTestJSON-1717540784-project-member] [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 763.446418] env[63088]: DEBUG oslo_concurrency.lockutils [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Acquiring lock "b159649b-ebd5-4c7d-9074-dc0e6395a947" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 763.446418] env[63088]: DEBUG oslo_concurrency.lockutils [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Lock "b159649b-ebd5-4c7d-9074-dc0e6395a947" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 763.475196] env[63088]: DEBUG nova.network.neutron [None req-3a58af89-5e90-4bc8-9f8f-c02b80d14311 tempest-InstanceActionsTestJSON-1717540784 tempest-InstanceActionsTestJSON-1717540784-project-member] [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 763.574071] env[63088]: DEBUG nova.network.neutron [None req-87ab5d07-9ebb-429d-8640-0197960e7d1a tempest-ServersNegativeTestMultiTenantJSON-2054426678 tempest-ServersNegativeTestMultiTenantJSON-2054426678-project-member] [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] Successfully created port: 5c2db8b3-531e-4a33-a138-03a6256495c1 {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 763.616551] env[63088]: DEBUG nova.compute.manager [None req-87ab5d07-9ebb-429d-8640-0197960e7d1a tempest-ServersNegativeTestMultiTenantJSON-2054426678 tempest-ServersNegativeTestMultiTenantJSON-2054426678-project-member] [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 763.783519] env[63088]: INFO nova.compute.manager [-] [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] Took 1.02 seconds to deallocate network for instance. [ 763.785595] env[63088]: DEBUG nova.compute.claims [None req-3edaa40f-e8ca-4f6f-a9f2-5744f490f65d tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] Aborting claim: {{(pid=63088) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 763.785692] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3edaa40f-e8ca-4f6f-a9f2-5744f490f65d tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 763.862957] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cda433f8-b6d6-4ee0-8d27-7ab160886124 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.870875] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7962487a-8032-4b8a-98a0-31f16392f48c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.905026] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91fd3e7d-aa01-42c2-853d-7bd45dfb139a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.909435] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37ea9f6d-dfc9-4496-bc9e-d3a1a61fd944 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.924054] env[63088]: DEBUG nova.compute.provider_tree [None req-aa39901a-ab16-4a54-a2f6-147052a0ed90 tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 763.976523] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3a58af89-5e90-4bc8-9f8f-c02b80d14311 tempest-InstanceActionsTestJSON-1717540784 tempest-InstanceActionsTestJSON-1717540784-project-member] Releasing lock "refresh_cache-7b5cc7b5-98b9-4344-b714-270dca798799" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 763.976908] env[63088]: DEBUG nova.compute.manager [None req-3a58af89-5e90-4bc8-9f8f-c02b80d14311 tempest-InstanceActionsTestJSON-1717540784 tempest-InstanceActionsTestJSON-1717540784-project-member] [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 763.977234] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-3a58af89-5e90-4bc8-9f8f-c02b80d14311 tempest-InstanceActionsTestJSON-1717540784 tempest-InstanceActionsTestJSON-1717540784-project-member] [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 763.977437] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8f6b46cc-37c7-48cc-816f-deed1dd947b9 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.985952] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-772b5639-05e4-432b-82b2-a0f9f3d35073 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.007674] env[63088]: WARNING nova.virt.vmwareapi.vmops [None req-3a58af89-5e90-4bc8-9f8f-c02b80d14311 tempest-InstanceActionsTestJSON-1717540784 tempest-InstanceActionsTestJSON-1717540784-project-member] [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7b5cc7b5-98b9-4344-b714-270dca798799 could not be found. [ 764.007883] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-3a58af89-5e90-4bc8-9f8f-c02b80d14311 tempest-InstanceActionsTestJSON-1717540784 tempest-InstanceActionsTestJSON-1717540784-project-member] [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 764.008072] env[63088]: INFO nova.compute.manager [None req-3a58af89-5e90-4bc8-9f8f-c02b80d14311 tempest-InstanceActionsTestJSON-1717540784 tempest-InstanceActionsTestJSON-1717540784-project-member] [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] Took 0.03 seconds to destroy the instance on the hypervisor. [ 764.008308] env[63088]: DEBUG oslo.service.loopingcall [None req-3a58af89-5e90-4bc8-9f8f-c02b80d14311 tempest-InstanceActionsTestJSON-1717540784 tempest-InstanceActionsTestJSON-1717540784-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 764.008519] env[63088]: DEBUG nova.compute.manager [-] [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 764.008613] env[63088]: DEBUG nova.network.neutron [-] [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 764.024018] env[63088]: DEBUG nova.network.neutron [-] [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 764.426834] env[63088]: DEBUG nova.scheduler.client.report [None req-aa39901a-ab16-4a54-a2f6-147052a0ed90 tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 764.526404] env[63088]: DEBUG nova.network.neutron [-] [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 764.620013] env[63088]: DEBUG nova.compute.manager [req-469d1005-ed74-40af-ac54-bb713bdf682e req-8235812b-7690-4e8c-94b3-7a128ca75977 service nova] [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] Received event network-changed-6904053f-859c-4b6f-80fc-0278d414bf42 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 764.620344] env[63088]: DEBUG nova.compute.manager [req-469d1005-ed74-40af-ac54-bb713bdf682e req-8235812b-7690-4e8c-94b3-7a128ca75977 service nova] [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] Refreshing instance network info cache due to event network-changed-6904053f-859c-4b6f-80fc-0278d414bf42. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 764.620463] env[63088]: DEBUG oslo_concurrency.lockutils [req-469d1005-ed74-40af-ac54-bb713bdf682e req-8235812b-7690-4e8c-94b3-7a128ca75977 service nova] Acquiring lock "refresh_cache-7b5cc7b5-98b9-4344-b714-270dca798799" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 764.620603] env[63088]: DEBUG oslo_concurrency.lockutils [req-469d1005-ed74-40af-ac54-bb713bdf682e req-8235812b-7690-4e8c-94b3-7a128ca75977 service nova] Acquired lock "refresh_cache-7b5cc7b5-98b9-4344-b714-270dca798799" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 764.620760] env[63088]: DEBUG nova.network.neutron [req-469d1005-ed74-40af-ac54-bb713bdf682e req-8235812b-7690-4e8c-94b3-7a128ca75977 service nova] [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] Refreshing network info cache for port 6904053f-859c-4b6f-80fc-0278d414bf42 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 764.627952] env[63088]: DEBUG nova.compute.manager [None req-87ab5d07-9ebb-429d-8640-0197960e7d1a tempest-ServersNegativeTestMultiTenantJSON-2054426678 tempest-ServersNegativeTestMultiTenantJSON-2054426678-project-member] [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 764.653776] env[63088]: DEBUG nova.virt.hardware [None req-87ab5d07-9ebb-429d-8640-0197960e7d1a tempest-ServersNegativeTestMultiTenantJSON-2054426678 tempest-ServersNegativeTestMultiTenantJSON-2054426678-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 764.654048] env[63088]: DEBUG nova.virt.hardware [None req-87ab5d07-9ebb-429d-8640-0197960e7d1a tempest-ServersNegativeTestMultiTenantJSON-2054426678 tempest-ServersNegativeTestMultiTenantJSON-2054426678-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 764.654211] env[63088]: DEBUG nova.virt.hardware [None req-87ab5d07-9ebb-429d-8640-0197960e7d1a tempest-ServersNegativeTestMultiTenantJSON-2054426678 tempest-ServersNegativeTestMultiTenantJSON-2054426678-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 764.654393] env[63088]: DEBUG nova.virt.hardware [None req-87ab5d07-9ebb-429d-8640-0197960e7d1a tempest-ServersNegativeTestMultiTenantJSON-2054426678 tempest-ServersNegativeTestMultiTenantJSON-2054426678-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 764.654539] env[63088]: DEBUG nova.virt.hardware [None req-87ab5d07-9ebb-429d-8640-0197960e7d1a tempest-ServersNegativeTestMultiTenantJSON-2054426678 tempest-ServersNegativeTestMultiTenantJSON-2054426678-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 764.654682] env[63088]: DEBUG nova.virt.hardware [None req-87ab5d07-9ebb-429d-8640-0197960e7d1a tempest-ServersNegativeTestMultiTenantJSON-2054426678 tempest-ServersNegativeTestMultiTenantJSON-2054426678-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 764.654889] env[63088]: DEBUG nova.virt.hardware [None req-87ab5d07-9ebb-429d-8640-0197960e7d1a tempest-ServersNegativeTestMultiTenantJSON-2054426678 tempest-ServersNegativeTestMultiTenantJSON-2054426678-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 764.655242] env[63088]: DEBUG nova.virt.hardware [None req-87ab5d07-9ebb-429d-8640-0197960e7d1a tempest-ServersNegativeTestMultiTenantJSON-2054426678 tempest-ServersNegativeTestMultiTenantJSON-2054426678-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 764.655495] env[63088]: DEBUG nova.virt.hardware [None req-87ab5d07-9ebb-429d-8640-0197960e7d1a tempest-ServersNegativeTestMultiTenantJSON-2054426678 tempest-ServersNegativeTestMultiTenantJSON-2054426678-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 764.655700] env[63088]: DEBUG nova.virt.hardware [None req-87ab5d07-9ebb-429d-8640-0197960e7d1a tempest-ServersNegativeTestMultiTenantJSON-2054426678 tempest-ServersNegativeTestMultiTenantJSON-2054426678-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 764.655907] env[63088]: DEBUG nova.virt.hardware [None req-87ab5d07-9ebb-429d-8640-0197960e7d1a tempest-ServersNegativeTestMultiTenantJSON-2054426678 tempest-ServersNegativeTestMultiTenantJSON-2054426678-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 764.656816] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b06f7b2f-6ea9-4a39-8c8d-17bf0ea89178 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.665320] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cb6e18a-d343-4b96-bdce-92a1710337d8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.937030] env[63088]: DEBUG oslo_concurrency.lockutils [None req-aa39901a-ab16-4a54-a2f6-147052a0ed90 tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.335s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 764.937574] env[63088]: DEBUG nova.compute.manager [None req-aa39901a-ab16-4a54-a2f6-147052a0ed90 tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] [instance: 43ded929-6890-4296-ae90-230f7c8b2595] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 764.940768] env[63088]: DEBUG oslo_concurrency.lockutils [None req-1c0c7b6c-8782-4a52-b8ee-b7fd4463c345 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.739s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 764.942452] env[63088]: INFO nova.compute.claims [None req-1c0c7b6c-8782-4a52-b8ee-b7fd4463c345 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 765.028771] env[63088]: INFO nova.compute.manager [-] [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] Took 1.02 seconds to deallocate network for instance. [ 765.031125] env[63088]: DEBUG nova.compute.claims [None req-3a58af89-5e90-4bc8-9f8f-c02b80d14311 tempest-InstanceActionsTestJSON-1717540784 tempest-InstanceActionsTestJSON-1717540784-project-member] [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] Aborting claim: {{(pid=63088) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 765.031302] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3a58af89-5e90-4bc8-9f8f-c02b80d14311 tempest-InstanceActionsTestJSON-1717540784 tempest-InstanceActionsTestJSON-1717540784-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 765.151779] env[63088]: DEBUG nova.network.neutron [req-469d1005-ed74-40af-ac54-bb713bdf682e req-8235812b-7690-4e8c-94b3-7a128ca75977 service nova] [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 765.234889] env[63088]: ERROR nova.compute.manager [None req-87ab5d07-9ebb-429d-8640-0197960e7d1a tempest-ServersNegativeTestMultiTenantJSON-2054426678 tempest-ServersNegativeTestMultiTenantJSON-2054426678-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5c2db8b3-531e-4a33-a138-03a6256495c1, please check neutron logs for more information. [ 765.234889] env[63088]: ERROR nova.compute.manager Traceback (most recent call last): [ 765.234889] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 765.234889] env[63088]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 765.234889] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 765.234889] env[63088]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 765.234889] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 765.234889] env[63088]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 765.234889] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 765.234889] env[63088]: ERROR nova.compute.manager self.force_reraise() [ 765.234889] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 765.234889] env[63088]: ERROR nova.compute.manager raise self.value [ 765.234889] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 765.234889] env[63088]: ERROR nova.compute.manager updated_port = self._update_port( [ 765.234889] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 765.234889] env[63088]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 765.235652] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 765.235652] env[63088]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 765.235652] env[63088]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5c2db8b3-531e-4a33-a138-03a6256495c1, please check neutron logs for more information. [ 765.235652] env[63088]: ERROR nova.compute.manager [ 765.235652] env[63088]: Traceback (most recent call last): [ 765.235652] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 765.235652] env[63088]: listener.cb(fileno) [ 765.235652] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 765.235652] env[63088]: result = function(*args, **kwargs) [ 765.235652] env[63088]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 765.235652] env[63088]: return func(*args, **kwargs) [ 765.235652] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 765.235652] env[63088]: raise e [ 765.235652] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 765.235652] env[63088]: nwinfo = self.network_api.allocate_for_instance( [ 765.235652] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 765.235652] env[63088]: created_port_ids = self._update_ports_for_instance( [ 765.235652] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 765.235652] env[63088]: with excutils.save_and_reraise_exception(): [ 765.235652] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 765.235652] env[63088]: self.force_reraise() [ 765.235652] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 765.235652] env[63088]: raise self.value [ 765.235652] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 765.235652] env[63088]: updated_port = self._update_port( [ 765.235652] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 765.235652] env[63088]: _ensure_no_port_binding_failure(port) [ 765.235652] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 765.235652] env[63088]: raise exception.PortBindingFailed(port_id=port['id']) [ 765.237108] env[63088]: nova.exception.PortBindingFailed: Binding failed for port 5c2db8b3-531e-4a33-a138-03a6256495c1, please check neutron logs for more information. [ 765.237108] env[63088]: Removing descriptor: 15 [ 765.237108] env[63088]: ERROR nova.compute.manager [None req-87ab5d07-9ebb-429d-8640-0197960e7d1a tempest-ServersNegativeTestMultiTenantJSON-2054426678 tempest-ServersNegativeTestMultiTenantJSON-2054426678-project-member] [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5c2db8b3-531e-4a33-a138-03a6256495c1, please check neutron logs for more information. [ 765.237108] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] Traceback (most recent call last): [ 765.237108] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 765.237108] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] yield resources [ 765.237108] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 765.237108] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] self.driver.spawn(context, instance, image_meta, [ 765.237108] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 765.237108] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] self._vmops.spawn(context, instance, image_meta, injected_files, [ 765.237108] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 765.237108] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] vm_ref = self.build_virtual_machine(instance, [ 765.237719] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 765.237719] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] vif_infos = vmwarevif.get_vif_info(self._session, [ 765.237719] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 765.237719] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] for vif in network_info: [ 765.237719] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 765.237719] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] return self._sync_wrapper(fn, *args, **kwargs) [ 765.237719] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 765.237719] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] self.wait() [ 765.237719] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 765.237719] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] self[:] = self._gt.wait() [ 765.237719] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 765.237719] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] return self._exit_event.wait() [ 765.237719] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 765.238337] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] result = hub.switch() [ 765.238337] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 765.238337] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] return self.greenlet.switch() [ 765.238337] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 765.238337] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] result = function(*args, **kwargs) [ 765.238337] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 765.238337] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] return func(*args, **kwargs) [ 765.238337] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 765.238337] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] raise e [ 765.238337] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 765.238337] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] nwinfo = self.network_api.allocate_for_instance( [ 765.238337] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 765.238337] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] created_port_ids = self._update_ports_for_instance( [ 765.239024] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 765.239024] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] with excutils.save_and_reraise_exception(): [ 765.239024] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 765.239024] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] self.force_reraise() [ 765.239024] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 765.239024] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] raise self.value [ 765.239024] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 765.239024] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] updated_port = self._update_port( [ 765.239024] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 765.239024] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] _ensure_no_port_binding_failure(port) [ 765.239024] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 765.239024] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] raise exception.PortBindingFailed(port_id=port['id']) [ 765.239804] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] nova.exception.PortBindingFailed: Binding failed for port 5c2db8b3-531e-4a33-a138-03a6256495c1, please check neutron logs for more information. [ 765.239804] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] [ 765.239804] env[63088]: INFO nova.compute.manager [None req-87ab5d07-9ebb-429d-8640-0197960e7d1a tempest-ServersNegativeTestMultiTenantJSON-2054426678 tempest-ServersNegativeTestMultiTenantJSON-2054426678-project-member] [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] Terminating instance [ 765.239804] env[63088]: DEBUG nova.network.neutron [req-469d1005-ed74-40af-ac54-bb713bdf682e req-8235812b-7690-4e8c-94b3-7a128ca75977 service nova] [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 765.239804] env[63088]: DEBUG oslo_concurrency.lockutils [None req-87ab5d07-9ebb-429d-8640-0197960e7d1a tempest-ServersNegativeTestMultiTenantJSON-2054426678 tempest-ServersNegativeTestMultiTenantJSON-2054426678-project-member] Acquiring lock "refresh_cache-d851a26b-48aa-4ffd-9c23-cea27543e0fa" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 765.239804] env[63088]: DEBUG oslo_concurrency.lockutils [None req-87ab5d07-9ebb-429d-8640-0197960e7d1a tempest-ServersNegativeTestMultiTenantJSON-2054426678 tempest-ServersNegativeTestMultiTenantJSON-2054426678-project-member] Acquired lock "refresh_cache-d851a26b-48aa-4ffd-9c23-cea27543e0fa" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 765.240065] env[63088]: DEBUG nova.network.neutron [None req-87ab5d07-9ebb-429d-8640-0197960e7d1a tempest-ServersNegativeTestMultiTenantJSON-2054426678 tempest-ServersNegativeTestMultiTenantJSON-2054426678-project-member] [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 765.442791] env[63088]: DEBUG nova.compute.utils [None req-aa39901a-ab16-4a54-a2f6-147052a0ed90 tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 765.445227] env[63088]: DEBUG nova.compute.manager [None req-aa39901a-ab16-4a54-a2f6-147052a0ed90 tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] [instance: 43ded929-6890-4296-ae90-230f7c8b2595] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 765.445227] env[63088]: DEBUG nova.network.neutron [None req-aa39901a-ab16-4a54-a2f6-147052a0ed90 tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] [instance: 43ded929-6890-4296-ae90-230f7c8b2595] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 765.504475] env[63088]: DEBUG nova.policy [None req-aa39901a-ab16-4a54-a2f6-147052a0ed90 tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '637ba82799ba4c09a3ee901704675860', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7cf8918ced834fdfaa1bf2350b666ce0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 765.742172] env[63088]: DEBUG oslo_concurrency.lockutils [req-469d1005-ed74-40af-ac54-bb713bdf682e req-8235812b-7690-4e8c-94b3-7a128ca75977 service nova] Releasing lock "refresh_cache-7b5cc7b5-98b9-4344-b714-270dca798799" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 765.742441] env[63088]: DEBUG nova.compute.manager [req-469d1005-ed74-40af-ac54-bb713bdf682e req-8235812b-7690-4e8c-94b3-7a128ca75977 service nova] [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] Received event network-vif-deleted-6904053f-859c-4b6f-80fc-0278d414bf42 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 765.763986] env[63088]: DEBUG nova.network.neutron [None req-87ab5d07-9ebb-429d-8640-0197960e7d1a tempest-ServersNegativeTestMultiTenantJSON-2054426678 tempest-ServersNegativeTestMultiTenantJSON-2054426678-project-member] [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 765.823075] env[63088]: DEBUG nova.network.neutron [None req-aa39901a-ab16-4a54-a2f6-147052a0ed90 tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] [instance: 43ded929-6890-4296-ae90-230f7c8b2595] Successfully created port: a0d4f41f-99bf-4a7a-9441-0835bc236922 {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 765.916709] env[63088]: DEBUG nova.network.neutron [None req-87ab5d07-9ebb-429d-8640-0197960e7d1a tempest-ServersNegativeTestMultiTenantJSON-2054426678 tempest-ServersNegativeTestMultiTenantJSON-2054426678-project-member] [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 765.951266] env[63088]: DEBUG nova.compute.manager [None req-aa39901a-ab16-4a54-a2f6-147052a0ed90 tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] [instance: 43ded929-6890-4296-ae90-230f7c8b2595] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 766.203398] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d27e4be-324b-47ea-8f88-dddd5a7271b1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.211177] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00e69a5e-bd66-4d46-ab29-31aaaebded53 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.240891] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3630baac-c12a-4d09-95ea-a098efd5aaf1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.248131] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16633c58-b0a8-4f33-a1d7-7a656a02691d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.262892] env[63088]: DEBUG nova.compute.provider_tree [None req-1c0c7b6c-8782-4a52-b8ee-b7fd4463c345 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 766.421740] env[63088]: DEBUG oslo_concurrency.lockutils [None req-87ab5d07-9ebb-429d-8640-0197960e7d1a tempest-ServersNegativeTestMultiTenantJSON-2054426678 tempest-ServersNegativeTestMultiTenantJSON-2054426678-project-member] Releasing lock "refresh_cache-d851a26b-48aa-4ffd-9c23-cea27543e0fa" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 766.421740] env[63088]: DEBUG nova.compute.manager [None req-87ab5d07-9ebb-429d-8640-0197960e7d1a tempest-ServersNegativeTestMultiTenantJSON-2054426678 tempest-ServersNegativeTestMultiTenantJSON-2054426678-project-member] [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 766.421740] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-87ab5d07-9ebb-429d-8640-0197960e7d1a tempest-ServersNegativeTestMultiTenantJSON-2054426678 tempest-ServersNegativeTestMultiTenantJSON-2054426678-project-member] [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 766.421740] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-141c1430-a5cd-4108-a600-1c6299478b0d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.430156] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbf3fae0-e34b-41d4-a61f-546428673ba9 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.473571] env[63088]: WARNING nova.virt.vmwareapi.vmops [None req-87ab5d07-9ebb-429d-8640-0197960e7d1a tempest-ServersNegativeTestMultiTenantJSON-2054426678 tempest-ServersNegativeTestMultiTenantJSON-2054426678-project-member] [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d851a26b-48aa-4ffd-9c23-cea27543e0fa could not be found. [ 766.473571] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-87ab5d07-9ebb-429d-8640-0197960e7d1a tempest-ServersNegativeTestMultiTenantJSON-2054426678 tempest-ServersNegativeTestMultiTenantJSON-2054426678-project-member] [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 766.473571] env[63088]: INFO nova.compute.manager [None req-87ab5d07-9ebb-429d-8640-0197960e7d1a tempest-ServersNegativeTestMultiTenantJSON-2054426678 tempest-ServersNegativeTestMultiTenantJSON-2054426678-project-member] [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] Took 0.05 seconds to destroy the instance on the hypervisor. [ 766.473571] env[63088]: DEBUG oslo.service.loopingcall [None req-87ab5d07-9ebb-429d-8640-0197960e7d1a tempest-ServersNegativeTestMultiTenantJSON-2054426678 tempest-ServersNegativeTestMultiTenantJSON-2054426678-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 766.473571] env[63088]: DEBUG nova.compute.manager [-] [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 766.473825] env[63088]: DEBUG nova.network.neutron [-] [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 766.504277] env[63088]: DEBUG nova.network.neutron [-] [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 766.569162] env[63088]: DEBUG nova.compute.manager [req-d0646030-2e0b-49c7-80b3-c89f7e283e49 req-30934081-c847-4042-8455-131549c64d90 service nova] [instance: 43ded929-6890-4296-ae90-230f7c8b2595] Received event network-changed-a0d4f41f-99bf-4a7a-9441-0835bc236922 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 766.569162] env[63088]: DEBUG nova.compute.manager [req-d0646030-2e0b-49c7-80b3-c89f7e283e49 req-30934081-c847-4042-8455-131549c64d90 service nova] [instance: 43ded929-6890-4296-ae90-230f7c8b2595] Refreshing instance network info cache due to event network-changed-a0d4f41f-99bf-4a7a-9441-0835bc236922. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 766.569162] env[63088]: DEBUG oslo_concurrency.lockutils [req-d0646030-2e0b-49c7-80b3-c89f7e283e49 req-30934081-c847-4042-8455-131549c64d90 service nova] Acquiring lock "refresh_cache-43ded929-6890-4296-ae90-230f7c8b2595" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 766.569162] env[63088]: DEBUG oslo_concurrency.lockutils [req-d0646030-2e0b-49c7-80b3-c89f7e283e49 req-30934081-c847-4042-8455-131549c64d90 service nova] Acquired lock "refresh_cache-43ded929-6890-4296-ae90-230f7c8b2595" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 766.569162] env[63088]: DEBUG nova.network.neutron [req-d0646030-2e0b-49c7-80b3-c89f7e283e49 req-30934081-c847-4042-8455-131549c64d90 service nova] [instance: 43ded929-6890-4296-ae90-230f7c8b2595] Refreshing network info cache for port a0d4f41f-99bf-4a7a-9441-0835bc236922 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 766.745105] env[63088]: ERROR nova.compute.manager [None req-aa39901a-ab16-4a54-a2f6-147052a0ed90 tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a0d4f41f-99bf-4a7a-9441-0835bc236922, please check neutron logs for more information. [ 766.745105] env[63088]: ERROR nova.compute.manager Traceback (most recent call last): [ 766.745105] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 766.745105] env[63088]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 766.745105] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 766.745105] env[63088]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 766.745105] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 766.745105] env[63088]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 766.745105] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 766.745105] env[63088]: ERROR nova.compute.manager self.force_reraise() [ 766.745105] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 766.745105] env[63088]: ERROR nova.compute.manager raise self.value [ 766.745105] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 766.745105] env[63088]: ERROR nova.compute.manager updated_port = self._update_port( [ 766.745105] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 766.745105] env[63088]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 766.745848] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 766.745848] env[63088]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 766.745848] env[63088]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a0d4f41f-99bf-4a7a-9441-0835bc236922, please check neutron logs for more information. [ 766.745848] env[63088]: ERROR nova.compute.manager [ 766.745848] env[63088]: Traceback (most recent call last): [ 766.745848] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 766.745848] env[63088]: listener.cb(fileno) [ 766.745848] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 766.745848] env[63088]: result = function(*args, **kwargs) [ 766.745848] env[63088]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 766.745848] env[63088]: return func(*args, **kwargs) [ 766.745848] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 766.745848] env[63088]: raise e [ 766.745848] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 766.745848] env[63088]: nwinfo = self.network_api.allocate_for_instance( [ 766.745848] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 766.745848] env[63088]: created_port_ids = self._update_ports_for_instance( [ 766.745848] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 766.745848] env[63088]: with excutils.save_and_reraise_exception(): [ 766.745848] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 766.745848] env[63088]: self.force_reraise() [ 766.745848] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 766.745848] env[63088]: raise self.value [ 766.745848] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 766.745848] env[63088]: updated_port = self._update_port( [ 766.745848] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 766.745848] env[63088]: _ensure_no_port_binding_failure(port) [ 766.745848] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 766.745848] env[63088]: raise exception.PortBindingFailed(port_id=port['id']) [ 766.746797] env[63088]: nova.exception.PortBindingFailed: Binding failed for port a0d4f41f-99bf-4a7a-9441-0835bc236922, please check neutron logs for more information. [ 766.746797] env[63088]: Removing descriptor: 17 [ 766.762200] env[63088]: DEBUG nova.compute.manager [req-7347a5d9-04f1-48d9-80de-12df82a38528 req-222e691a-ee99-4b3c-a80d-141edf4bf691 service nova] [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] Received event network-changed-5c2db8b3-531e-4a33-a138-03a6256495c1 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 766.762200] env[63088]: DEBUG nova.compute.manager [req-7347a5d9-04f1-48d9-80de-12df82a38528 req-222e691a-ee99-4b3c-a80d-141edf4bf691 service nova] [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] Refreshing instance network info cache due to event network-changed-5c2db8b3-531e-4a33-a138-03a6256495c1. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 766.762200] env[63088]: DEBUG oslo_concurrency.lockutils [req-7347a5d9-04f1-48d9-80de-12df82a38528 req-222e691a-ee99-4b3c-a80d-141edf4bf691 service nova] Acquiring lock "refresh_cache-d851a26b-48aa-4ffd-9c23-cea27543e0fa" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 766.762200] env[63088]: DEBUG oslo_concurrency.lockutils [req-7347a5d9-04f1-48d9-80de-12df82a38528 req-222e691a-ee99-4b3c-a80d-141edf4bf691 service nova] Acquired lock "refresh_cache-d851a26b-48aa-4ffd-9c23-cea27543e0fa" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 766.762200] env[63088]: DEBUG nova.network.neutron [req-7347a5d9-04f1-48d9-80de-12df82a38528 req-222e691a-ee99-4b3c-a80d-141edf4bf691 service nova] [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] Refreshing network info cache for port 5c2db8b3-531e-4a33-a138-03a6256495c1 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 766.768943] env[63088]: DEBUG nova.scheduler.client.report [None req-1c0c7b6c-8782-4a52-b8ee-b7fd4463c345 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 766.963419] env[63088]: DEBUG nova.compute.manager [None req-aa39901a-ab16-4a54-a2f6-147052a0ed90 tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] [instance: 43ded929-6890-4296-ae90-230f7c8b2595] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 766.989020] env[63088]: DEBUG nova.virt.hardware [None req-aa39901a-ab16-4a54-a2f6-147052a0ed90 tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 766.989177] env[63088]: DEBUG nova.virt.hardware [None req-aa39901a-ab16-4a54-a2f6-147052a0ed90 tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 766.989347] env[63088]: DEBUG nova.virt.hardware [None req-aa39901a-ab16-4a54-a2f6-147052a0ed90 tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 766.989510] env[63088]: DEBUG nova.virt.hardware [None req-aa39901a-ab16-4a54-a2f6-147052a0ed90 tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 766.989657] env[63088]: DEBUG nova.virt.hardware [None req-aa39901a-ab16-4a54-a2f6-147052a0ed90 tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 766.989800] env[63088]: DEBUG nova.virt.hardware [None req-aa39901a-ab16-4a54-a2f6-147052a0ed90 tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 766.990014] env[63088]: DEBUG nova.virt.hardware [None req-aa39901a-ab16-4a54-a2f6-147052a0ed90 tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 766.990177] env[63088]: DEBUG nova.virt.hardware [None req-aa39901a-ab16-4a54-a2f6-147052a0ed90 tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 766.990347] env[63088]: DEBUG nova.virt.hardware [None req-aa39901a-ab16-4a54-a2f6-147052a0ed90 tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 766.990539] env[63088]: DEBUG nova.virt.hardware [None req-aa39901a-ab16-4a54-a2f6-147052a0ed90 tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 766.990717] env[63088]: DEBUG nova.virt.hardware [None req-aa39901a-ab16-4a54-a2f6-147052a0ed90 tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 766.991581] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-166280a0-6fad-4088-bd54-6455adef0c39 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.999915] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf2ec4f8-0497-4f5d-b116-f7c65daa22dd {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.014253] env[63088]: DEBUG nova.network.neutron [-] [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 767.016506] env[63088]: ERROR nova.compute.manager [None req-aa39901a-ab16-4a54-a2f6-147052a0ed90 tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] [instance: 43ded929-6890-4296-ae90-230f7c8b2595] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a0d4f41f-99bf-4a7a-9441-0835bc236922, please check neutron logs for more information. [ 767.016506] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] Traceback (most recent call last): [ 767.016506] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 767.016506] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] yield resources [ 767.016506] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 767.016506] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] self.driver.spawn(context, instance, image_meta, [ 767.016506] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 767.016506] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] self._vmops.spawn(context, instance, image_meta, injected_files, [ 767.016506] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 767.016506] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] vm_ref = self.build_virtual_machine(instance, [ 767.016506] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 767.016891] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] vif_infos = vmwarevif.get_vif_info(self._session, [ 767.016891] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 767.016891] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] for vif in network_info: [ 767.016891] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 767.016891] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] return self._sync_wrapper(fn, *args, **kwargs) [ 767.016891] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 767.016891] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] self.wait() [ 767.016891] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 767.016891] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] self[:] = self._gt.wait() [ 767.016891] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 767.016891] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] return self._exit_event.wait() [ 767.016891] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 767.016891] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] current.throw(*self._exc) [ 767.017430] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 767.017430] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] result = function(*args, **kwargs) [ 767.017430] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 767.017430] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] return func(*args, **kwargs) [ 767.017430] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 767.017430] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] raise e [ 767.017430] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 767.017430] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] nwinfo = self.network_api.allocate_for_instance( [ 767.017430] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 767.017430] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] created_port_ids = self._update_ports_for_instance( [ 767.017430] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 767.017430] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] with excutils.save_and_reraise_exception(): [ 767.017430] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 767.017858] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] self.force_reraise() [ 767.017858] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 767.017858] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] raise self.value [ 767.017858] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 767.017858] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] updated_port = self._update_port( [ 767.017858] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 767.017858] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] _ensure_no_port_binding_failure(port) [ 767.017858] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 767.017858] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] raise exception.PortBindingFailed(port_id=port['id']) [ 767.017858] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] nova.exception.PortBindingFailed: Binding failed for port a0d4f41f-99bf-4a7a-9441-0835bc236922, please check neutron logs for more information. [ 767.017858] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] [ 767.017858] env[63088]: INFO nova.compute.manager [None req-aa39901a-ab16-4a54-a2f6-147052a0ed90 tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] [instance: 43ded929-6890-4296-ae90-230f7c8b2595] Terminating instance [ 767.018577] env[63088]: DEBUG oslo_concurrency.lockutils [None req-aa39901a-ab16-4a54-a2f6-147052a0ed90 tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] Acquiring lock "refresh_cache-43ded929-6890-4296-ae90-230f7c8b2595" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 767.087413] env[63088]: DEBUG nova.network.neutron [req-d0646030-2e0b-49c7-80b3-c89f7e283e49 req-30934081-c847-4042-8455-131549c64d90 service nova] [instance: 43ded929-6890-4296-ae90-230f7c8b2595] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 767.173862] env[63088]: DEBUG nova.network.neutron [req-d0646030-2e0b-49c7-80b3-c89f7e283e49 req-30934081-c847-4042-8455-131549c64d90 service nova] [instance: 43ded929-6890-4296-ae90-230f7c8b2595] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 767.275300] env[63088]: DEBUG oslo_concurrency.lockutils [None req-1c0c7b6c-8782-4a52-b8ee-b7fd4463c345 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.334s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 767.275846] env[63088]: DEBUG nova.compute.manager [None req-1c0c7b6c-8782-4a52-b8ee-b7fd4463c345 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 767.278455] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d012c7de-73d3-432e-8cc9-9d2f0456777d tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.660s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 767.285578] env[63088]: DEBUG nova.network.neutron [req-7347a5d9-04f1-48d9-80de-12df82a38528 req-222e691a-ee99-4b3c-a80d-141edf4bf691 service nova] [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 767.374162] env[63088]: DEBUG nova.network.neutron [req-7347a5d9-04f1-48d9-80de-12df82a38528 req-222e691a-ee99-4b3c-a80d-141edf4bf691 service nova] [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 767.516427] env[63088]: INFO nova.compute.manager [-] [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] Took 1.04 seconds to deallocate network for instance. [ 767.518881] env[63088]: DEBUG nova.compute.claims [None req-87ab5d07-9ebb-429d-8640-0197960e7d1a tempest-ServersNegativeTestMultiTenantJSON-2054426678 tempest-ServersNegativeTestMultiTenantJSON-2054426678-project-member] [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] Aborting claim: {{(pid=63088) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 767.519073] env[63088]: DEBUG oslo_concurrency.lockutils [None req-87ab5d07-9ebb-429d-8640-0197960e7d1a tempest-ServersNegativeTestMultiTenantJSON-2054426678 tempest-ServersNegativeTestMultiTenantJSON-2054426678-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 767.676901] env[63088]: DEBUG oslo_concurrency.lockutils [req-d0646030-2e0b-49c7-80b3-c89f7e283e49 req-30934081-c847-4042-8455-131549c64d90 service nova] Releasing lock "refresh_cache-43ded929-6890-4296-ae90-230f7c8b2595" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 767.677088] env[63088]: DEBUG oslo_concurrency.lockutils [None req-aa39901a-ab16-4a54-a2f6-147052a0ed90 tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] Acquired lock "refresh_cache-43ded929-6890-4296-ae90-230f7c8b2595" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 767.677336] env[63088]: DEBUG nova.network.neutron [None req-aa39901a-ab16-4a54-a2f6-147052a0ed90 tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] [instance: 43ded929-6890-4296-ae90-230f7c8b2595] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 767.784014] env[63088]: DEBUG nova.compute.utils [None req-1c0c7b6c-8782-4a52-b8ee-b7fd4463c345 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 767.789229] env[63088]: DEBUG nova.compute.manager [None req-1c0c7b6c-8782-4a52-b8ee-b7fd4463c345 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 767.789462] env[63088]: DEBUG nova.network.neutron [None req-1c0c7b6c-8782-4a52-b8ee-b7fd4463c345 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 767.833046] env[63088]: DEBUG nova.policy [None req-1c0c7b6c-8782-4a52-b8ee-b7fd4463c345 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a6f91fc8922c42f985f6f6338dab0708', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7bff1382b9694df08133c88a5fe783a5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 767.877682] env[63088]: DEBUG oslo_concurrency.lockutils [req-7347a5d9-04f1-48d9-80de-12df82a38528 req-222e691a-ee99-4b3c-a80d-141edf4bf691 service nova] Releasing lock "refresh_cache-d851a26b-48aa-4ffd-9c23-cea27543e0fa" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 767.877682] env[63088]: DEBUG nova.compute.manager [req-7347a5d9-04f1-48d9-80de-12df82a38528 req-222e691a-ee99-4b3c-a80d-141edf4bf691 service nova] [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] Received event network-vif-deleted-5c2db8b3-531e-4a33-a138-03a6256495c1 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 768.044882] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3420741-3edb-4179-97ad-b9f80fbf9302 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.052555] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6921274-e51b-42af-9fad-8e815b55f75f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.082071] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-746ede6f-637e-4f58-9720-8dc6103ba7cb {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.088747] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75958853-1d54-4440-b488-d8c887dcd871 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.101347] env[63088]: DEBUG nova.compute.provider_tree [None req-d012c7de-73d3-432e-8cc9-9d2f0456777d tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 768.202266] env[63088]: DEBUG nova.network.neutron [None req-aa39901a-ab16-4a54-a2f6-147052a0ed90 tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] [instance: 43ded929-6890-4296-ae90-230f7c8b2595] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 768.251914] env[63088]: DEBUG nova.network.neutron [None req-1c0c7b6c-8782-4a52-b8ee-b7fd4463c345 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] Successfully created port: 09226e11-0d58-4ff6-adfe-80141e1092c2 {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 768.289888] env[63088]: DEBUG nova.compute.manager [None req-1c0c7b6c-8782-4a52-b8ee-b7fd4463c345 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 768.299072] env[63088]: DEBUG nova.network.neutron [None req-aa39901a-ab16-4a54-a2f6-147052a0ed90 tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] [instance: 43ded929-6890-4296-ae90-230f7c8b2595] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 768.597084] env[63088]: DEBUG nova.compute.manager [req-570e2aa3-6b09-46c2-b58d-4734271b29de req-f4492e0a-7c44-4b55-9eb0-abba80c249a9 service nova] [instance: 43ded929-6890-4296-ae90-230f7c8b2595] Received event network-vif-deleted-a0d4f41f-99bf-4a7a-9441-0835bc236922 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 768.606211] env[63088]: DEBUG nova.scheduler.client.report [None req-d012c7de-73d3-432e-8cc9-9d2f0456777d tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 768.804418] env[63088]: DEBUG oslo_concurrency.lockutils [None req-aa39901a-ab16-4a54-a2f6-147052a0ed90 tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] Releasing lock "refresh_cache-43ded929-6890-4296-ae90-230f7c8b2595" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 768.804418] env[63088]: DEBUG nova.compute.manager [None req-aa39901a-ab16-4a54-a2f6-147052a0ed90 tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] [instance: 43ded929-6890-4296-ae90-230f7c8b2595] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 768.804418] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-aa39901a-ab16-4a54-a2f6-147052a0ed90 tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] [instance: 43ded929-6890-4296-ae90-230f7c8b2595] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 768.804418] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a1ad5e59-7abf-443a-bce3-0e549dc7787e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.813866] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70c62fa4-8de3-45b1-9121-fbb30a25b9eb {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.839685] env[63088]: WARNING nova.virt.vmwareapi.vmops [None req-aa39901a-ab16-4a54-a2f6-147052a0ed90 tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] [instance: 43ded929-6890-4296-ae90-230f7c8b2595] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 43ded929-6890-4296-ae90-230f7c8b2595 could not be found. [ 768.839685] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-aa39901a-ab16-4a54-a2f6-147052a0ed90 tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] [instance: 43ded929-6890-4296-ae90-230f7c8b2595] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 768.839685] env[63088]: INFO nova.compute.manager [None req-aa39901a-ab16-4a54-a2f6-147052a0ed90 tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] [instance: 43ded929-6890-4296-ae90-230f7c8b2595] Took 0.04 seconds to destroy the instance on the hypervisor. [ 768.839685] env[63088]: DEBUG oslo.service.loopingcall [None req-aa39901a-ab16-4a54-a2f6-147052a0ed90 tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 768.839685] env[63088]: DEBUG nova.compute.manager [-] [instance: 43ded929-6890-4296-ae90-230f7c8b2595] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 768.839685] env[63088]: DEBUG nova.network.neutron [-] [instance: 43ded929-6890-4296-ae90-230f7c8b2595] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 769.019094] env[63088]: DEBUG nova.network.neutron [-] [instance: 43ded929-6890-4296-ae90-230f7c8b2595] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 769.108558] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d012c7de-73d3-432e-8cc9-9d2f0456777d tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.830s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 769.109845] env[63088]: ERROR nova.compute.manager [None req-d012c7de-73d3-432e-8cc9-9d2f0456777d tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 170f1cb1-c3ef-431f-8ef9-de984c0f19d1, please check neutron logs for more information. [ 769.109845] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] Traceback (most recent call last): [ 769.109845] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 769.109845] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] self.driver.spawn(context, instance, image_meta, [ 769.109845] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 769.109845] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] self._vmops.spawn(context, instance, image_meta, injected_files, [ 769.109845] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 769.109845] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] vm_ref = self.build_virtual_machine(instance, [ 769.109845] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 769.109845] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] vif_infos = vmwarevif.get_vif_info(self._session, [ 769.109845] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 769.112115] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] for vif in network_info: [ 769.112115] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 769.112115] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] return self._sync_wrapper(fn, *args, **kwargs) [ 769.112115] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 769.112115] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] self.wait() [ 769.112115] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 769.112115] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] self[:] = self._gt.wait() [ 769.112115] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 769.112115] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] return self._exit_event.wait() [ 769.112115] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 769.112115] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] result = hub.switch() [ 769.112115] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 769.112115] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] return self.greenlet.switch() [ 769.112509] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 769.112509] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] result = function(*args, **kwargs) [ 769.112509] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 769.112509] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] return func(*args, **kwargs) [ 769.112509] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 769.112509] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] raise e [ 769.112509] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 769.112509] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] nwinfo = self.network_api.allocate_for_instance( [ 769.112509] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 769.112509] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] created_port_ids = self._update_ports_for_instance( [ 769.112509] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 769.112509] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] with excutils.save_and_reraise_exception(): [ 769.112509] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 769.112872] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] self.force_reraise() [ 769.112872] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 769.112872] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] raise self.value [ 769.112872] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 769.112872] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] updated_port = self._update_port( [ 769.112872] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 769.112872] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] _ensure_no_port_binding_failure(port) [ 769.112872] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 769.112872] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] raise exception.PortBindingFailed(port_id=port['id']) [ 769.112872] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] nova.exception.PortBindingFailed: Binding failed for port 170f1cb1-c3ef-431f-8ef9-de984c0f19d1, please check neutron logs for more information. [ 769.112872] env[63088]: ERROR nova.compute.manager [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] [ 769.113200] env[63088]: DEBUG nova.compute.utils [None req-d012c7de-73d3-432e-8cc9-9d2f0456777d tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] Binding failed for port 170f1cb1-c3ef-431f-8ef9-de984c0f19d1, please check neutron logs for more information. {{(pid=63088) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 769.114015] env[63088]: DEBUG nova.compute.manager [None req-d012c7de-73d3-432e-8cc9-9d2f0456777d tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] Build of instance 641eed47-54f6-4c08-a8fd-f06bc5f7fb56 was re-scheduled: Binding failed for port 170f1cb1-c3ef-431f-8ef9-de984c0f19d1, please check neutron logs for more information. {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 769.114496] env[63088]: DEBUG nova.compute.manager [None req-d012c7de-73d3-432e-8cc9-9d2f0456777d tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] Unplugging VIFs for instance {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 769.114762] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d012c7de-73d3-432e-8cc9-9d2f0456777d tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Acquiring lock "refresh_cache-641eed47-54f6-4c08-a8fd-f06bc5f7fb56" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 769.114947] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d012c7de-73d3-432e-8cc9-9d2f0456777d tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Acquired lock "refresh_cache-641eed47-54f6-4c08-a8fd-f06bc5f7fb56" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 769.115167] env[63088]: DEBUG nova.network.neutron [None req-d012c7de-73d3-432e-8cc9-9d2f0456777d tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 769.116263] env[63088]: DEBUG oslo_concurrency.lockutils [None req-96c00a75-0383-4506-a4c6-d540aa18aaef tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.487s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 769.117953] env[63088]: INFO nova.compute.claims [None req-96c00a75-0383-4506-a4c6-d540aa18aaef tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 769.259875] env[63088]: ERROR nova.compute.manager [None req-1c0c7b6c-8782-4a52-b8ee-b7fd4463c345 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 09226e11-0d58-4ff6-adfe-80141e1092c2, please check neutron logs for more information. [ 769.259875] env[63088]: ERROR nova.compute.manager Traceback (most recent call last): [ 769.259875] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 769.259875] env[63088]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 769.259875] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 769.259875] env[63088]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 769.259875] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 769.259875] env[63088]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 769.259875] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 769.259875] env[63088]: ERROR nova.compute.manager self.force_reraise() [ 769.259875] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 769.259875] env[63088]: ERROR nova.compute.manager raise self.value [ 769.259875] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 769.259875] env[63088]: ERROR nova.compute.manager updated_port = self._update_port( [ 769.259875] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 769.259875] env[63088]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 769.260421] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 769.260421] env[63088]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 769.260421] env[63088]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 09226e11-0d58-4ff6-adfe-80141e1092c2, please check neutron logs for more information. [ 769.260421] env[63088]: ERROR nova.compute.manager [ 769.260421] env[63088]: Traceback (most recent call last): [ 769.260421] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 769.260421] env[63088]: listener.cb(fileno) [ 769.260421] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 769.260421] env[63088]: result = function(*args, **kwargs) [ 769.260421] env[63088]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 769.260421] env[63088]: return func(*args, **kwargs) [ 769.260421] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 769.260421] env[63088]: raise e [ 769.260421] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 769.260421] env[63088]: nwinfo = self.network_api.allocate_for_instance( [ 769.260421] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 769.260421] env[63088]: created_port_ids = self._update_ports_for_instance( [ 769.260421] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 769.260421] env[63088]: with excutils.save_and_reraise_exception(): [ 769.260421] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 769.260421] env[63088]: self.force_reraise() [ 769.260421] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 769.260421] env[63088]: raise self.value [ 769.260421] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 769.260421] env[63088]: updated_port = self._update_port( [ 769.260421] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 769.260421] env[63088]: _ensure_no_port_binding_failure(port) [ 769.260421] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 769.260421] env[63088]: raise exception.PortBindingFailed(port_id=port['id']) [ 769.262493] env[63088]: nova.exception.PortBindingFailed: Binding failed for port 09226e11-0d58-4ff6-adfe-80141e1092c2, please check neutron logs for more information. [ 769.262493] env[63088]: Removing descriptor: 17 [ 769.301166] env[63088]: DEBUG nova.compute.manager [None req-1c0c7b6c-8782-4a52-b8ee-b7fd4463c345 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 769.332312] env[63088]: DEBUG nova.virt.hardware [None req-1c0c7b6c-8782-4a52-b8ee-b7fd4463c345 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 769.332647] env[63088]: DEBUG nova.virt.hardware [None req-1c0c7b6c-8782-4a52-b8ee-b7fd4463c345 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 769.332806] env[63088]: DEBUG nova.virt.hardware [None req-1c0c7b6c-8782-4a52-b8ee-b7fd4463c345 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 769.332986] env[63088]: DEBUG nova.virt.hardware [None req-1c0c7b6c-8782-4a52-b8ee-b7fd4463c345 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 769.334122] env[63088]: DEBUG nova.virt.hardware [None req-1c0c7b6c-8782-4a52-b8ee-b7fd4463c345 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 769.334300] env[63088]: DEBUG nova.virt.hardware [None req-1c0c7b6c-8782-4a52-b8ee-b7fd4463c345 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 769.334513] env[63088]: DEBUG nova.virt.hardware [None req-1c0c7b6c-8782-4a52-b8ee-b7fd4463c345 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 769.334678] env[63088]: DEBUG nova.virt.hardware [None req-1c0c7b6c-8782-4a52-b8ee-b7fd4463c345 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 769.334844] env[63088]: DEBUG nova.virt.hardware [None req-1c0c7b6c-8782-4a52-b8ee-b7fd4463c345 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 769.335009] env[63088]: DEBUG nova.virt.hardware [None req-1c0c7b6c-8782-4a52-b8ee-b7fd4463c345 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 769.335192] env[63088]: DEBUG nova.virt.hardware [None req-1c0c7b6c-8782-4a52-b8ee-b7fd4463c345 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 769.336053] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16fb6743-8357-4e10-854a-6973126cc920 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.344382] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a5aaaee-bcd3-4f10-94f2-9afa028a5b32 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.358272] env[63088]: ERROR nova.compute.manager [None req-1c0c7b6c-8782-4a52-b8ee-b7fd4463c345 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 09226e11-0d58-4ff6-adfe-80141e1092c2, please check neutron logs for more information. [ 769.358272] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] Traceback (most recent call last): [ 769.358272] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 769.358272] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] yield resources [ 769.358272] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 769.358272] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] self.driver.spawn(context, instance, image_meta, [ 769.358272] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 769.358272] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 769.358272] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 769.358272] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] vm_ref = self.build_virtual_machine(instance, [ 769.358272] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 769.358696] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] vif_infos = vmwarevif.get_vif_info(self._session, [ 769.358696] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 769.358696] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] for vif in network_info: [ 769.358696] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 769.358696] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] return self._sync_wrapper(fn, *args, **kwargs) [ 769.358696] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 769.358696] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] self.wait() [ 769.358696] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 769.358696] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] self[:] = self._gt.wait() [ 769.358696] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 769.358696] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] return self._exit_event.wait() [ 769.358696] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 769.358696] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] current.throw(*self._exc) [ 769.359138] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 769.359138] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] result = function(*args, **kwargs) [ 769.359138] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 769.359138] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] return func(*args, **kwargs) [ 769.359138] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 769.359138] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] raise e [ 769.359138] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 769.359138] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] nwinfo = self.network_api.allocate_for_instance( [ 769.359138] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 769.359138] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] created_port_ids = self._update_ports_for_instance( [ 769.359138] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 769.359138] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] with excutils.save_and_reraise_exception(): [ 769.359138] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 769.359595] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] self.force_reraise() [ 769.359595] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 769.359595] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] raise self.value [ 769.359595] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 769.359595] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] updated_port = self._update_port( [ 769.359595] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 769.359595] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] _ensure_no_port_binding_failure(port) [ 769.359595] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 769.359595] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] raise exception.PortBindingFailed(port_id=port['id']) [ 769.359595] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] nova.exception.PortBindingFailed: Binding failed for port 09226e11-0d58-4ff6-adfe-80141e1092c2, please check neutron logs for more information. [ 769.359595] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] [ 769.359595] env[63088]: INFO nova.compute.manager [None req-1c0c7b6c-8782-4a52-b8ee-b7fd4463c345 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] Terminating instance [ 769.360608] env[63088]: DEBUG oslo_concurrency.lockutils [None req-1c0c7b6c-8782-4a52-b8ee-b7fd4463c345 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Acquiring lock "refresh_cache-fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 769.360743] env[63088]: DEBUG oslo_concurrency.lockutils [None req-1c0c7b6c-8782-4a52-b8ee-b7fd4463c345 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Acquired lock "refresh_cache-fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 769.360905] env[63088]: DEBUG nova.network.neutron [None req-1c0c7b6c-8782-4a52-b8ee-b7fd4463c345 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 769.520806] env[63088]: DEBUG nova.network.neutron [-] [instance: 43ded929-6890-4296-ae90-230f7c8b2595] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 769.641391] env[63088]: DEBUG nova.network.neutron [None req-d012c7de-73d3-432e-8cc9-9d2f0456777d tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 769.738894] env[63088]: DEBUG nova.network.neutron [None req-d012c7de-73d3-432e-8cc9-9d2f0456777d tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 769.891248] env[63088]: DEBUG nova.network.neutron [None req-1c0c7b6c-8782-4a52-b8ee-b7fd4463c345 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 769.992832] env[63088]: DEBUG nova.network.neutron [None req-1c0c7b6c-8782-4a52-b8ee-b7fd4463c345 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 770.023663] env[63088]: INFO nova.compute.manager [-] [instance: 43ded929-6890-4296-ae90-230f7c8b2595] Took 1.18 seconds to deallocate network for instance. [ 770.025374] env[63088]: DEBUG nova.compute.claims [None req-aa39901a-ab16-4a54-a2f6-147052a0ed90 tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] [instance: 43ded929-6890-4296-ae90-230f7c8b2595] Aborting claim: {{(pid=63088) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 770.025569] env[63088]: DEBUG oslo_concurrency.lockutils [None req-aa39901a-ab16-4a54-a2f6-147052a0ed90 tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 770.241397] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d012c7de-73d3-432e-8cc9-9d2f0456777d tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Releasing lock "refresh_cache-641eed47-54f6-4c08-a8fd-f06bc5f7fb56" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 770.241954] env[63088]: DEBUG nova.compute.manager [None req-d012c7de-73d3-432e-8cc9-9d2f0456777d tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 770.241954] env[63088]: DEBUG nova.compute.manager [None req-d012c7de-73d3-432e-8cc9-9d2f0456777d tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 770.242073] env[63088]: DEBUG nova.network.neutron [None req-d012c7de-73d3-432e-8cc9-9d2f0456777d tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 770.263601] env[63088]: DEBUG nova.network.neutron [None req-d012c7de-73d3-432e-8cc9-9d2f0456777d tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 770.382178] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0f7c059-973f-4dbd-971c-0be450f3b8a9 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.391038] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0837d425-f08b-4f46-9937-320269b80c35 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.422160] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38c858ec-254b-496d-990c-392bdac52d31 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.428920] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8f23e28-8556-476e-b010-472ac95cd42f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.443021] env[63088]: DEBUG nova.compute.provider_tree [None req-96c00a75-0383-4506-a4c6-d540aa18aaef tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 770.495362] env[63088]: DEBUG oslo_concurrency.lockutils [None req-1c0c7b6c-8782-4a52-b8ee-b7fd4463c345 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Releasing lock "refresh_cache-fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 770.496116] env[63088]: DEBUG nova.compute.manager [None req-1c0c7b6c-8782-4a52-b8ee-b7fd4463c345 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 770.496116] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-1c0c7b6c-8782-4a52-b8ee-b7fd4463c345 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 770.496279] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-52bf47ec-50af-4d11-922c-f4b5769218d9 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.505215] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db4d3f5e-0f42-4ccb-bac1-1d35a699bc6c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.527989] env[63088]: WARNING nova.virt.vmwareapi.vmops [None req-1c0c7b6c-8782-4a52-b8ee-b7fd4463c345 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3 could not be found. [ 770.528233] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-1c0c7b6c-8782-4a52-b8ee-b7fd4463c345 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 770.528414] env[63088]: INFO nova.compute.manager [None req-1c0c7b6c-8782-4a52-b8ee-b7fd4463c345 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] Took 0.03 seconds to destroy the instance on the hypervisor. [ 770.528654] env[63088]: DEBUG oslo.service.loopingcall [None req-1c0c7b6c-8782-4a52-b8ee-b7fd4463c345 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 770.528879] env[63088]: DEBUG nova.compute.manager [-] [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 770.528944] env[63088]: DEBUG nova.network.neutron [-] [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 770.551278] env[63088]: DEBUG nova.network.neutron [-] [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 770.646128] env[63088]: DEBUG nova.compute.manager [req-7d56518f-f56f-400b-97e9-793020e05724 req-95f89eac-f3ef-4c99-a6cc-7c89f4349c12 service nova] [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] Received event network-changed-09226e11-0d58-4ff6-adfe-80141e1092c2 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 770.646534] env[63088]: DEBUG nova.compute.manager [req-7d56518f-f56f-400b-97e9-793020e05724 req-95f89eac-f3ef-4c99-a6cc-7c89f4349c12 service nova] [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] Refreshing instance network info cache due to event network-changed-09226e11-0d58-4ff6-adfe-80141e1092c2. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 770.648025] env[63088]: DEBUG oslo_concurrency.lockutils [req-7d56518f-f56f-400b-97e9-793020e05724 req-95f89eac-f3ef-4c99-a6cc-7c89f4349c12 service nova] Acquiring lock "refresh_cache-fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 770.648025] env[63088]: DEBUG oslo_concurrency.lockutils [req-7d56518f-f56f-400b-97e9-793020e05724 req-95f89eac-f3ef-4c99-a6cc-7c89f4349c12 service nova] Acquired lock "refresh_cache-fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 770.648025] env[63088]: DEBUG nova.network.neutron [req-7d56518f-f56f-400b-97e9-793020e05724 req-95f89eac-f3ef-4c99-a6cc-7c89f4349c12 service nova] [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] Refreshing network info cache for port 09226e11-0d58-4ff6-adfe-80141e1092c2 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 770.766493] env[63088]: DEBUG nova.network.neutron [None req-d012c7de-73d3-432e-8cc9-9d2f0456777d tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 770.945939] env[63088]: DEBUG nova.scheduler.client.report [None req-96c00a75-0383-4506-a4c6-d540aa18aaef tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 771.053786] env[63088]: DEBUG nova.network.neutron [-] [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 771.165708] env[63088]: DEBUG nova.network.neutron [req-7d56518f-f56f-400b-97e9-793020e05724 req-95f89eac-f3ef-4c99-a6cc-7c89f4349c12 service nova] [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 771.260159] env[63088]: DEBUG nova.network.neutron [req-7d56518f-f56f-400b-97e9-793020e05724 req-95f89eac-f3ef-4c99-a6cc-7c89f4349c12 service nova] [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 771.268726] env[63088]: INFO nova.compute.manager [None req-d012c7de-73d3-432e-8cc9-9d2f0456777d tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] [instance: 641eed47-54f6-4c08-a8fd-f06bc5f7fb56] Took 1.03 seconds to deallocate network for instance. [ 771.451319] env[63088]: DEBUG oslo_concurrency.lockutils [None req-96c00a75-0383-4506-a4c6-d540aa18aaef tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.335s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 771.451827] env[63088]: DEBUG nova.compute.manager [None req-96c00a75-0383-4506-a4c6-d540aa18aaef tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 771.454940] env[63088]: DEBUG oslo_concurrency.lockutils [None req-02aa52fc-420f-46cd-9829-a68800c429e2 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.700s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 771.456583] env[63088]: INFO nova.compute.claims [None req-02aa52fc-420f-46cd-9829-a68800c429e2 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 771.556476] env[63088]: INFO nova.compute.manager [-] [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] Took 1.03 seconds to deallocate network for instance. [ 771.558638] env[63088]: DEBUG nova.compute.claims [None req-1c0c7b6c-8782-4a52-b8ee-b7fd4463c345 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] Aborting claim: {{(pid=63088) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 771.558809] env[63088]: DEBUG oslo_concurrency.lockutils [None req-1c0c7b6c-8782-4a52-b8ee-b7fd4463c345 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 771.764417] env[63088]: DEBUG oslo_concurrency.lockutils [req-7d56518f-f56f-400b-97e9-793020e05724 req-95f89eac-f3ef-4c99-a6cc-7c89f4349c12 service nova] Releasing lock "refresh_cache-fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 771.764713] env[63088]: DEBUG nova.compute.manager [req-7d56518f-f56f-400b-97e9-793020e05724 req-95f89eac-f3ef-4c99-a6cc-7c89f4349c12 service nova] [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] Received event network-vif-deleted-09226e11-0d58-4ff6-adfe-80141e1092c2 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 771.964513] env[63088]: DEBUG nova.compute.utils [None req-96c00a75-0383-4506-a4c6-d540aa18aaef tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 771.966021] env[63088]: DEBUG nova.compute.manager [None req-96c00a75-0383-4506-a4c6-d540aa18aaef tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 771.966347] env[63088]: DEBUG nova.network.neutron [None req-96c00a75-0383-4506-a4c6-d540aa18aaef tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 772.012836] env[63088]: DEBUG nova.policy [None req-96c00a75-0383-4506-a4c6-d540aa18aaef tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0d4202df2fce486a96fea968d0189398', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7ed1e04b27ee464eb9a72866d8327352', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 772.018836] env[63088]: DEBUG oslo_concurrency.lockutils [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Acquiring lock "34ce411f-40c8-446e-b685-cd3ce07663dd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 772.019062] env[63088]: DEBUG oslo_concurrency.lockutils [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Lock "34ce411f-40c8-446e-b685-cd3ce07663dd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 772.303806] env[63088]: INFO nova.scheduler.client.report [None req-d012c7de-73d3-432e-8cc9-9d2f0456777d tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Deleted allocations for instance 641eed47-54f6-4c08-a8fd-f06bc5f7fb56 [ 772.313722] env[63088]: DEBUG nova.network.neutron [None req-96c00a75-0383-4506-a4c6-d540aa18aaef tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] Successfully created port: 5d70ba36-7221-4886-8ae1-21223c696acf {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 772.470461] env[63088]: DEBUG nova.compute.manager [None req-96c00a75-0383-4506-a4c6-d540aa18aaef tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 772.740793] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3225be1-377e-4d89-ae5e-a7312277a2e3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.749042] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d0e2f33-7504-464e-8c3e-c6d61b8db9f7 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.779050] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-908a13c5-0238-418b-b224-728587648a3c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.786207] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29564a08-368e-433e-9e2f-8e8d25c0e4ec {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.800137] env[63088]: DEBUG nova.compute.provider_tree [None req-02aa52fc-420f-46cd-9829-a68800c429e2 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 772.809443] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d012c7de-73d3-432e-8cc9-9d2f0456777d tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Lock "641eed47-54f6-4c08-a8fd-f06bc5f7fb56" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 137.666s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 773.183874] env[63088]: DEBUG nova.compute.manager [req-f47a4e89-94d3-4b28-a4dd-1a2241757e4f req-7712e44d-235d-4b25-8dad-47d0d4142d56 service nova] [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] Received event network-changed-5d70ba36-7221-4886-8ae1-21223c696acf {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 773.184210] env[63088]: DEBUG nova.compute.manager [req-f47a4e89-94d3-4b28-a4dd-1a2241757e4f req-7712e44d-235d-4b25-8dad-47d0d4142d56 service nova] [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] Refreshing instance network info cache due to event network-changed-5d70ba36-7221-4886-8ae1-21223c696acf. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 773.184499] env[63088]: DEBUG oslo_concurrency.lockutils [req-f47a4e89-94d3-4b28-a4dd-1a2241757e4f req-7712e44d-235d-4b25-8dad-47d0d4142d56 service nova] Acquiring lock "refresh_cache-87361d1c-e9cd-4177-bf3d-5d7551192073" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 773.184656] env[63088]: DEBUG oslo_concurrency.lockutils [req-f47a4e89-94d3-4b28-a4dd-1a2241757e4f req-7712e44d-235d-4b25-8dad-47d0d4142d56 service nova] Acquired lock "refresh_cache-87361d1c-e9cd-4177-bf3d-5d7551192073" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 773.184854] env[63088]: DEBUG nova.network.neutron [req-f47a4e89-94d3-4b28-a4dd-1a2241757e4f req-7712e44d-235d-4b25-8dad-47d0d4142d56 service nova] [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] Refreshing network info cache for port 5d70ba36-7221-4886-8ae1-21223c696acf {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 773.303839] env[63088]: DEBUG nova.scheduler.client.report [None req-02aa52fc-420f-46cd-9829-a68800c429e2 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 773.313023] env[63088]: DEBUG nova.compute.manager [None req-b5c531a2-a229-4998-9a9b-9ebb1e7afff2 tempest-ServerGroupTestJSON-1534054659 tempest-ServerGroupTestJSON-1534054659-project-member] [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 773.481714] env[63088]: ERROR nova.compute.manager [None req-96c00a75-0383-4506-a4c6-d540aa18aaef tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5d70ba36-7221-4886-8ae1-21223c696acf, please check neutron logs for more information. [ 773.481714] env[63088]: ERROR nova.compute.manager Traceback (most recent call last): [ 773.481714] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 773.481714] env[63088]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 773.481714] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 773.481714] env[63088]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 773.481714] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 773.481714] env[63088]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 773.481714] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 773.481714] env[63088]: ERROR nova.compute.manager self.force_reraise() [ 773.481714] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 773.481714] env[63088]: ERROR nova.compute.manager raise self.value [ 773.481714] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 773.481714] env[63088]: ERROR nova.compute.manager updated_port = self._update_port( [ 773.481714] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 773.481714] env[63088]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 773.482272] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 773.482272] env[63088]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 773.482272] env[63088]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5d70ba36-7221-4886-8ae1-21223c696acf, please check neutron logs for more information. [ 773.482272] env[63088]: ERROR nova.compute.manager [ 773.482272] env[63088]: Traceback (most recent call last): [ 773.482272] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 773.482272] env[63088]: listener.cb(fileno) [ 773.482272] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 773.482272] env[63088]: result = function(*args, **kwargs) [ 773.482272] env[63088]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 773.482272] env[63088]: return func(*args, **kwargs) [ 773.482272] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 773.482272] env[63088]: raise e [ 773.482272] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 773.482272] env[63088]: nwinfo = self.network_api.allocate_for_instance( [ 773.482272] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 773.482272] env[63088]: created_port_ids = self._update_ports_for_instance( [ 773.482272] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 773.482272] env[63088]: with excutils.save_and_reraise_exception(): [ 773.482272] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 773.482272] env[63088]: self.force_reraise() [ 773.482272] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 773.482272] env[63088]: raise self.value [ 773.482272] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 773.482272] env[63088]: updated_port = self._update_port( [ 773.482272] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 773.482272] env[63088]: _ensure_no_port_binding_failure(port) [ 773.482272] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 773.482272] env[63088]: raise exception.PortBindingFailed(port_id=port['id']) [ 773.483142] env[63088]: nova.exception.PortBindingFailed: Binding failed for port 5d70ba36-7221-4886-8ae1-21223c696acf, please check neutron logs for more information. [ 773.483142] env[63088]: Removing descriptor: 17 [ 773.484629] env[63088]: DEBUG nova.compute.manager [None req-96c00a75-0383-4506-a4c6-d540aa18aaef tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 773.511661] env[63088]: DEBUG nova.virt.hardware [None req-96c00a75-0383-4506-a4c6-d540aa18aaef tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 773.511936] env[63088]: DEBUG nova.virt.hardware [None req-96c00a75-0383-4506-a4c6-d540aa18aaef tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 773.512128] env[63088]: DEBUG nova.virt.hardware [None req-96c00a75-0383-4506-a4c6-d540aa18aaef tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 773.512313] env[63088]: DEBUG nova.virt.hardware [None req-96c00a75-0383-4506-a4c6-d540aa18aaef tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 773.512483] env[63088]: DEBUG nova.virt.hardware [None req-96c00a75-0383-4506-a4c6-d540aa18aaef tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 773.512601] env[63088]: DEBUG nova.virt.hardware [None req-96c00a75-0383-4506-a4c6-d540aa18aaef tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 773.512797] env[63088]: DEBUG nova.virt.hardware [None req-96c00a75-0383-4506-a4c6-d540aa18aaef tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 773.512952] env[63088]: DEBUG nova.virt.hardware [None req-96c00a75-0383-4506-a4c6-d540aa18aaef tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 773.513134] env[63088]: DEBUG nova.virt.hardware [None req-96c00a75-0383-4506-a4c6-d540aa18aaef tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 773.513982] env[63088]: DEBUG nova.virt.hardware [None req-96c00a75-0383-4506-a4c6-d540aa18aaef tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 773.513982] env[63088]: DEBUG nova.virt.hardware [None req-96c00a75-0383-4506-a4c6-d540aa18aaef tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 773.514350] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ced3db8-0214-4738-8cc7-4ff605757d46 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.522189] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66913720-bf24-443f-be3f-440f15bd4043 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.535756] env[63088]: ERROR nova.compute.manager [None req-96c00a75-0383-4506-a4c6-d540aa18aaef tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5d70ba36-7221-4886-8ae1-21223c696acf, please check neutron logs for more information. [ 773.535756] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] Traceback (most recent call last): [ 773.535756] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 773.535756] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] yield resources [ 773.535756] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 773.535756] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] self.driver.spawn(context, instance, image_meta, [ 773.535756] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 773.535756] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] self._vmops.spawn(context, instance, image_meta, injected_files, [ 773.535756] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 773.535756] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] vm_ref = self.build_virtual_machine(instance, [ 773.535756] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 773.536195] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] vif_infos = vmwarevif.get_vif_info(self._session, [ 773.536195] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 773.536195] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] for vif in network_info: [ 773.536195] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 773.536195] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] return self._sync_wrapper(fn, *args, **kwargs) [ 773.536195] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 773.536195] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] self.wait() [ 773.536195] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 773.536195] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] self[:] = self._gt.wait() [ 773.536195] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 773.536195] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] return self._exit_event.wait() [ 773.536195] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 773.536195] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] current.throw(*self._exc) [ 773.536617] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 773.536617] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] result = function(*args, **kwargs) [ 773.536617] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 773.536617] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] return func(*args, **kwargs) [ 773.536617] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 773.536617] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] raise e [ 773.536617] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 773.536617] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] nwinfo = self.network_api.allocate_for_instance( [ 773.536617] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 773.536617] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] created_port_ids = self._update_ports_for_instance( [ 773.536617] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 773.536617] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] with excutils.save_and_reraise_exception(): [ 773.536617] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 773.537064] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] self.force_reraise() [ 773.537064] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 773.537064] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] raise self.value [ 773.537064] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 773.537064] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] updated_port = self._update_port( [ 773.537064] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 773.537064] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] _ensure_no_port_binding_failure(port) [ 773.537064] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 773.537064] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] raise exception.PortBindingFailed(port_id=port['id']) [ 773.537064] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] nova.exception.PortBindingFailed: Binding failed for port 5d70ba36-7221-4886-8ae1-21223c696acf, please check neutron logs for more information. [ 773.537064] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] [ 773.537064] env[63088]: INFO nova.compute.manager [None req-96c00a75-0383-4506-a4c6-d540aa18aaef tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] Terminating instance [ 773.539508] env[63088]: DEBUG oslo_concurrency.lockutils [None req-96c00a75-0383-4506-a4c6-d540aa18aaef tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Acquiring lock "refresh_cache-87361d1c-e9cd-4177-bf3d-5d7551192073" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 773.703099] env[63088]: DEBUG nova.network.neutron [req-f47a4e89-94d3-4b28-a4dd-1a2241757e4f req-7712e44d-235d-4b25-8dad-47d0d4142d56 service nova] [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 773.790439] env[63088]: DEBUG nova.network.neutron [req-f47a4e89-94d3-4b28-a4dd-1a2241757e4f req-7712e44d-235d-4b25-8dad-47d0d4142d56 service nova] [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 773.809022] env[63088]: DEBUG oslo_concurrency.lockutils [None req-02aa52fc-420f-46cd-9829-a68800c429e2 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.354s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 773.809184] env[63088]: DEBUG nova.compute.manager [None req-02aa52fc-420f-46cd-9829-a68800c429e2 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 773.813019] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bce00a2e-7fb0-44a9-a66e-261b7c3d5384 tempest-ServerActionsV293TestJSON-1088890325 tempest-ServerActionsV293TestJSON-1088890325-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.116s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 773.848945] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b5c531a2-a229-4998-9a9b-9ebb1e7afff2 tempest-ServerGroupTestJSON-1534054659 tempest-ServerGroupTestJSON-1534054659-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 774.293478] env[63088]: DEBUG oslo_concurrency.lockutils [req-f47a4e89-94d3-4b28-a4dd-1a2241757e4f req-7712e44d-235d-4b25-8dad-47d0d4142d56 service nova] Releasing lock "refresh_cache-87361d1c-e9cd-4177-bf3d-5d7551192073" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 774.293914] env[63088]: DEBUG oslo_concurrency.lockutils [None req-96c00a75-0383-4506-a4c6-d540aa18aaef tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Acquired lock "refresh_cache-87361d1c-e9cd-4177-bf3d-5d7551192073" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 774.294127] env[63088]: DEBUG nova.network.neutron [None req-96c00a75-0383-4506-a4c6-d540aa18aaef tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 774.313492] env[63088]: DEBUG nova.compute.utils [None req-02aa52fc-420f-46cd-9829-a68800c429e2 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 774.314784] env[63088]: DEBUG nova.compute.manager [None req-02aa52fc-420f-46cd-9829-a68800c429e2 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 774.314956] env[63088]: DEBUG nova.network.neutron [None req-02aa52fc-420f-46cd-9829-a68800c429e2 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 774.389094] env[63088]: DEBUG nova.policy [None req-02aa52fc-420f-46cd-9829-a68800c429e2 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9b77ca5186e5437a862d644b2075fe8b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '982309216b714182b5304c4d15808b9e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 774.417383] env[63088]: DEBUG oslo_concurrency.lockutils [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Acquiring lock "1d421bf0-925d-408c-a612-b659da92799f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 774.417635] env[63088]: DEBUG oslo_concurrency.lockutils [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Lock "1d421bf0-925d-408c-a612-b659da92799f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 774.557238] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-722f885e-39a9-446d-9fc2-298dd90d338a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.565416] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d178192d-8ecc-4b76-b879-a27291814fcc {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.599731] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04bac7e9-5b76-4dbc-9459-9e9e65b29103 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.604975] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-333012ad-200b-4d03-8f60-ca0f19a6965c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.618327] env[63088]: DEBUG nova.compute.provider_tree [None req-bce00a2e-7fb0-44a9-a66e-261b7c3d5384 tempest-ServerActionsV293TestJSON-1088890325 tempest-ServerActionsV293TestJSON-1088890325-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 774.789518] env[63088]: DEBUG nova.network.neutron [None req-02aa52fc-420f-46cd-9829-a68800c429e2 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] Successfully created port: 7876ef7c-e226-4a28-ac65-0d3647d7d12c {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 774.811654] env[63088]: DEBUG nova.network.neutron [None req-96c00a75-0383-4506-a4c6-d540aa18aaef tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 774.817815] env[63088]: DEBUG nova.compute.manager [None req-02aa52fc-420f-46cd-9829-a68800c429e2 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 774.926174] env[63088]: DEBUG nova.network.neutron [None req-96c00a75-0383-4506-a4c6-d540aa18aaef tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 775.121631] env[63088]: DEBUG nova.scheduler.client.report [None req-bce00a2e-7fb0-44a9-a66e-261b7c3d5384 tempest-ServerActionsV293TestJSON-1088890325 tempest-ServerActionsV293TestJSON-1088890325-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 775.225876] env[63088]: DEBUG nova.compute.manager [req-0dadc975-405d-41df-b0a4-654e29d8d29b req-aa5825a3-068c-4f82-9902-fea76f32f4b9 service nova] [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] Received event network-vif-deleted-5d70ba36-7221-4886-8ae1-21223c696acf {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 775.429061] env[63088]: DEBUG oslo_concurrency.lockutils [None req-96c00a75-0383-4506-a4c6-d540aa18aaef tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Releasing lock "refresh_cache-87361d1c-e9cd-4177-bf3d-5d7551192073" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 775.429556] env[63088]: DEBUG nova.compute.manager [None req-96c00a75-0383-4506-a4c6-d540aa18aaef tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 775.429759] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-96c00a75-0383-4506-a4c6-d540aa18aaef tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 775.430052] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3788fa64-7d36-4511-aec4-867f3817a4f4 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.438680] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc7621d8-3963-4a0a-9236-96ec6813ad5d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.458964] env[63088]: WARNING nova.virt.vmwareapi.vmops [None req-96c00a75-0383-4506-a4c6-d540aa18aaef tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 87361d1c-e9cd-4177-bf3d-5d7551192073 could not be found. [ 775.459175] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-96c00a75-0383-4506-a4c6-d540aa18aaef tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 775.459407] env[63088]: INFO nova.compute.manager [None req-96c00a75-0383-4506-a4c6-d540aa18aaef tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] Took 0.03 seconds to destroy the instance on the hypervisor. [ 775.459674] env[63088]: DEBUG oslo.service.loopingcall [None req-96c00a75-0383-4506-a4c6-d540aa18aaef tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 775.459891] env[63088]: DEBUG nova.compute.manager [-] [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 775.459975] env[63088]: DEBUG nova.network.neutron [-] [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 775.474504] env[63088]: DEBUG nova.network.neutron [-] [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 775.626695] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bce00a2e-7fb0-44a9-a66e-261b7c3d5384 tempest-ServerActionsV293TestJSON-1088890325 tempest-ServerActionsV293TestJSON-1088890325-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.814s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 775.627593] env[63088]: ERROR nova.compute.manager [None req-bce00a2e-7fb0-44a9-a66e-261b7c3d5384 tempest-ServerActionsV293TestJSON-1088890325 tempest-ServerActionsV293TestJSON-1088890325-project-member] [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 685b39f8-6e13-43dc-9dfa-6be110e1ac12, please check neutron logs for more information. [ 775.627593] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] Traceback (most recent call last): [ 775.627593] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 775.627593] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] self.driver.spawn(context, instance, image_meta, [ 775.627593] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 775.627593] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 775.627593] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 775.627593] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] vm_ref = self.build_virtual_machine(instance, [ 775.627593] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 775.627593] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] vif_infos = vmwarevif.get_vif_info(self._session, [ 775.627593] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 775.627950] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] for vif in network_info: [ 775.627950] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 775.627950] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] return self._sync_wrapper(fn, *args, **kwargs) [ 775.627950] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 775.627950] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] self.wait() [ 775.627950] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 775.627950] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] self[:] = self._gt.wait() [ 775.627950] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 775.627950] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] return self._exit_event.wait() [ 775.627950] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 775.627950] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] current.throw(*self._exc) [ 775.627950] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 775.627950] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] result = function(*args, **kwargs) [ 775.628363] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 775.628363] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] return func(*args, **kwargs) [ 775.628363] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 775.628363] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] raise e [ 775.628363] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 775.628363] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] nwinfo = self.network_api.allocate_for_instance( [ 775.628363] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 775.628363] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] created_port_ids = self._update_ports_for_instance( [ 775.628363] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 775.628363] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] with excutils.save_and_reraise_exception(): [ 775.628363] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 775.628363] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] self.force_reraise() [ 775.628363] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 775.628723] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] raise self.value [ 775.628723] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 775.628723] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] updated_port = self._update_port( [ 775.628723] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 775.628723] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] _ensure_no_port_binding_failure(port) [ 775.628723] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 775.628723] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] raise exception.PortBindingFailed(port_id=port['id']) [ 775.628723] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] nova.exception.PortBindingFailed: Binding failed for port 685b39f8-6e13-43dc-9dfa-6be110e1ac12, please check neutron logs for more information. [ 775.628723] env[63088]: ERROR nova.compute.manager [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] [ 775.628723] env[63088]: DEBUG nova.compute.utils [None req-bce00a2e-7fb0-44a9-a66e-261b7c3d5384 tempest-ServerActionsV293TestJSON-1088890325 tempest-ServerActionsV293TestJSON-1088890325-project-member] [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] Binding failed for port 685b39f8-6e13-43dc-9dfa-6be110e1ac12, please check neutron logs for more information. {{(pid=63088) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 775.631138] env[63088]: DEBUG oslo_concurrency.lockutils [None req-1874938e-086f-4c3c-9049-55ea0246719c tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.525s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 775.634046] env[63088]: DEBUG nova.compute.manager [None req-bce00a2e-7fb0-44a9-a66e-261b7c3d5384 tempest-ServerActionsV293TestJSON-1088890325 tempest-ServerActionsV293TestJSON-1088890325-project-member] [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] Build of instance b82c89ad-f7d6-4c04-a48e-acdab29ea5e7 was re-scheduled: Binding failed for port 685b39f8-6e13-43dc-9dfa-6be110e1ac12, please check neutron logs for more information. {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 775.634484] env[63088]: DEBUG nova.compute.manager [None req-bce00a2e-7fb0-44a9-a66e-261b7c3d5384 tempest-ServerActionsV293TestJSON-1088890325 tempest-ServerActionsV293TestJSON-1088890325-project-member] [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] Unplugging VIFs for instance {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 775.634716] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bce00a2e-7fb0-44a9-a66e-261b7c3d5384 tempest-ServerActionsV293TestJSON-1088890325 tempest-ServerActionsV293TestJSON-1088890325-project-member] Acquiring lock "refresh_cache-b82c89ad-f7d6-4c04-a48e-acdab29ea5e7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 775.634878] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bce00a2e-7fb0-44a9-a66e-261b7c3d5384 tempest-ServerActionsV293TestJSON-1088890325 tempest-ServerActionsV293TestJSON-1088890325-project-member] Acquired lock "refresh_cache-b82c89ad-f7d6-4c04-a48e-acdab29ea5e7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 775.635048] env[63088]: DEBUG nova.network.neutron [None req-bce00a2e-7fb0-44a9-a66e-261b7c3d5384 tempest-ServerActionsV293TestJSON-1088890325 tempest-ServerActionsV293TestJSON-1088890325-project-member] [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 775.725422] env[63088]: ERROR nova.compute.manager [None req-02aa52fc-420f-46cd-9829-a68800c429e2 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 7876ef7c-e226-4a28-ac65-0d3647d7d12c, please check neutron logs for more information. [ 775.725422] env[63088]: ERROR nova.compute.manager Traceback (most recent call last): [ 775.725422] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 775.725422] env[63088]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 775.725422] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 775.725422] env[63088]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 775.725422] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 775.725422] env[63088]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 775.725422] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 775.725422] env[63088]: ERROR nova.compute.manager self.force_reraise() [ 775.725422] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 775.725422] env[63088]: ERROR nova.compute.manager raise self.value [ 775.725422] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 775.725422] env[63088]: ERROR nova.compute.manager updated_port = self._update_port( [ 775.725422] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 775.725422] env[63088]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 775.725970] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 775.725970] env[63088]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 775.725970] env[63088]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 7876ef7c-e226-4a28-ac65-0d3647d7d12c, please check neutron logs for more information. [ 775.725970] env[63088]: ERROR nova.compute.manager [ 775.725970] env[63088]: Traceback (most recent call last): [ 775.725970] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 775.725970] env[63088]: listener.cb(fileno) [ 775.725970] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 775.725970] env[63088]: result = function(*args, **kwargs) [ 775.725970] env[63088]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 775.725970] env[63088]: return func(*args, **kwargs) [ 775.725970] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 775.725970] env[63088]: raise e [ 775.725970] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 775.725970] env[63088]: nwinfo = self.network_api.allocate_for_instance( [ 775.725970] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 775.725970] env[63088]: created_port_ids = self._update_ports_for_instance( [ 775.725970] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 775.725970] env[63088]: with excutils.save_and_reraise_exception(): [ 775.725970] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 775.725970] env[63088]: self.force_reraise() [ 775.725970] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 775.725970] env[63088]: raise self.value [ 775.725970] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 775.725970] env[63088]: updated_port = self._update_port( [ 775.725970] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 775.725970] env[63088]: _ensure_no_port_binding_failure(port) [ 775.725970] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 775.725970] env[63088]: raise exception.PortBindingFailed(port_id=port['id']) [ 775.726940] env[63088]: nova.exception.PortBindingFailed: Binding failed for port 7876ef7c-e226-4a28-ac65-0d3647d7d12c, please check neutron logs for more information. [ 775.726940] env[63088]: Removing descriptor: 17 [ 775.831776] env[63088]: DEBUG nova.compute.manager [None req-02aa52fc-420f-46cd-9829-a68800c429e2 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 775.857451] env[63088]: DEBUG nova.virt.hardware [None req-02aa52fc-420f-46cd-9829-a68800c429e2 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 775.857699] env[63088]: DEBUG nova.virt.hardware [None req-02aa52fc-420f-46cd-9829-a68800c429e2 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 775.857854] env[63088]: DEBUG nova.virt.hardware [None req-02aa52fc-420f-46cd-9829-a68800c429e2 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 775.858041] env[63088]: DEBUG nova.virt.hardware [None req-02aa52fc-420f-46cd-9829-a68800c429e2 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 775.858193] env[63088]: DEBUG nova.virt.hardware [None req-02aa52fc-420f-46cd-9829-a68800c429e2 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 775.858354] env[63088]: DEBUG nova.virt.hardware [None req-02aa52fc-420f-46cd-9829-a68800c429e2 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 775.858565] env[63088]: DEBUG nova.virt.hardware [None req-02aa52fc-420f-46cd-9829-a68800c429e2 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 775.858720] env[63088]: DEBUG nova.virt.hardware [None req-02aa52fc-420f-46cd-9829-a68800c429e2 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 775.858882] env[63088]: DEBUG nova.virt.hardware [None req-02aa52fc-420f-46cd-9829-a68800c429e2 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 775.859054] env[63088]: DEBUG nova.virt.hardware [None req-02aa52fc-420f-46cd-9829-a68800c429e2 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 775.859228] env[63088]: DEBUG nova.virt.hardware [None req-02aa52fc-420f-46cd-9829-a68800c429e2 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 775.860183] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27fc5c1d-992e-459e-ba1e-3bb8940127fe {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.868411] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb1431f4-6ee9-401e-9cfe-ea56bcc5a745 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.881925] env[63088]: ERROR nova.compute.manager [None req-02aa52fc-420f-46cd-9829-a68800c429e2 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 7876ef7c-e226-4a28-ac65-0d3647d7d12c, please check neutron logs for more information. [ 775.881925] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] Traceback (most recent call last): [ 775.881925] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 775.881925] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] yield resources [ 775.881925] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 775.881925] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] self.driver.spawn(context, instance, image_meta, [ 775.881925] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 775.881925] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 775.881925] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 775.881925] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] vm_ref = self.build_virtual_machine(instance, [ 775.881925] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 775.882521] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] vif_infos = vmwarevif.get_vif_info(self._session, [ 775.882521] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 775.882521] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] for vif in network_info: [ 775.882521] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 775.882521] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] return self._sync_wrapper(fn, *args, **kwargs) [ 775.882521] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 775.882521] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] self.wait() [ 775.882521] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 775.882521] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] self[:] = self._gt.wait() [ 775.882521] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 775.882521] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] return self._exit_event.wait() [ 775.882521] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 775.882521] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] current.throw(*self._exc) [ 775.883107] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 775.883107] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] result = function(*args, **kwargs) [ 775.883107] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 775.883107] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] return func(*args, **kwargs) [ 775.883107] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 775.883107] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] raise e [ 775.883107] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 775.883107] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] nwinfo = self.network_api.allocate_for_instance( [ 775.883107] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 775.883107] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] created_port_ids = self._update_ports_for_instance( [ 775.883107] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 775.883107] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] with excutils.save_and_reraise_exception(): [ 775.883107] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 775.883556] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] self.force_reraise() [ 775.883556] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 775.883556] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] raise self.value [ 775.883556] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 775.883556] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] updated_port = self._update_port( [ 775.883556] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 775.883556] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] _ensure_no_port_binding_failure(port) [ 775.883556] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 775.883556] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] raise exception.PortBindingFailed(port_id=port['id']) [ 775.883556] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] nova.exception.PortBindingFailed: Binding failed for port 7876ef7c-e226-4a28-ac65-0d3647d7d12c, please check neutron logs for more information. [ 775.883556] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] [ 775.883556] env[63088]: INFO nova.compute.manager [None req-02aa52fc-420f-46cd-9829-a68800c429e2 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] Terminating instance [ 775.884237] env[63088]: DEBUG oslo_concurrency.lockutils [None req-02aa52fc-420f-46cd-9829-a68800c429e2 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Acquiring lock "refresh_cache-e6a72b08-38b6-44ae-8911-6b39d50ee1c3" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 775.884395] env[63088]: DEBUG oslo_concurrency.lockutils [None req-02aa52fc-420f-46cd-9829-a68800c429e2 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Acquired lock "refresh_cache-e6a72b08-38b6-44ae-8911-6b39d50ee1c3" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 775.884559] env[63088]: DEBUG nova.network.neutron [None req-02aa52fc-420f-46cd-9829-a68800c429e2 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 775.976422] env[63088]: DEBUG nova.network.neutron [-] [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 776.157168] env[63088]: DEBUG nova.network.neutron [None req-bce00a2e-7fb0-44a9-a66e-261b7c3d5384 tempest-ServerActionsV293TestJSON-1088890325 tempest-ServerActionsV293TestJSON-1088890325-project-member] [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 776.253119] env[63088]: DEBUG nova.network.neutron [None req-bce00a2e-7fb0-44a9-a66e-261b7c3d5384 tempest-ServerActionsV293TestJSON-1088890325 tempest-ServerActionsV293TestJSON-1088890325-project-member] [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 776.362265] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32e105e8-859d-4aa6-aac5-bf48e7c831f8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.370028] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3694fb0-1fe7-4cfa-b181-ef72d740e3b0 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.401775] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a39bdb1-2f5f-4769-b080-5be08a109c5e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.408908] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79d49b10-1987-4dac-9e00-d02ae9798cfd {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.421780] env[63088]: DEBUG nova.compute.provider_tree [None req-1874938e-086f-4c3c-9049-55ea0246719c tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 776.423463] env[63088]: DEBUG nova.network.neutron [None req-02aa52fc-420f-46cd-9829-a68800c429e2 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 776.476967] env[63088]: DEBUG nova.network.neutron [None req-02aa52fc-420f-46cd-9829-a68800c429e2 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 776.479636] env[63088]: INFO nova.compute.manager [-] [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] Took 1.02 seconds to deallocate network for instance. [ 776.481590] env[63088]: DEBUG nova.compute.claims [None req-96c00a75-0383-4506-a4c6-d540aa18aaef tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] Aborting claim: {{(pid=63088) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 776.481761] env[63088]: DEBUG oslo_concurrency.lockutils [None req-96c00a75-0383-4506-a4c6-d540aa18aaef tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 776.756164] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bce00a2e-7fb0-44a9-a66e-261b7c3d5384 tempest-ServerActionsV293TestJSON-1088890325 tempest-ServerActionsV293TestJSON-1088890325-project-member] Releasing lock "refresh_cache-b82c89ad-f7d6-4c04-a48e-acdab29ea5e7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 776.756421] env[63088]: DEBUG nova.compute.manager [None req-bce00a2e-7fb0-44a9-a66e-261b7c3d5384 tempest-ServerActionsV293TestJSON-1088890325 tempest-ServerActionsV293TestJSON-1088890325-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 776.756604] env[63088]: DEBUG nova.compute.manager [None req-bce00a2e-7fb0-44a9-a66e-261b7c3d5384 tempest-ServerActionsV293TestJSON-1088890325 tempest-ServerActionsV293TestJSON-1088890325-project-member] [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 776.756773] env[63088]: DEBUG nova.network.neutron [None req-bce00a2e-7fb0-44a9-a66e-261b7c3d5384 tempest-ServerActionsV293TestJSON-1088890325 tempest-ServerActionsV293TestJSON-1088890325-project-member] [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 776.771178] env[63088]: DEBUG nova.network.neutron [None req-bce00a2e-7fb0-44a9-a66e-261b7c3d5384 tempest-ServerActionsV293TestJSON-1088890325 tempest-ServerActionsV293TestJSON-1088890325-project-member] [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 776.926632] env[63088]: DEBUG nova.scheduler.client.report [None req-1874938e-086f-4c3c-9049-55ea0246719c tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 776.980954] env[63088]: DEBUG oslo_concurrency.lockutils [None req-02aa52fc-420f-46cd-9829-a68800c429e2 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Releasing lock "refresh_cache-e6a72b08-38b6-44ae-8911-6b39d50ee1c3" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 776.981335] env[63088]: DEBUG nova.compute.manager [None req-02aa52fc-420f-46cd-9829-a68800c429e2 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 776.981476] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-02aa52fc-420f-46cd-9829-a68800c429e2 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 776.981788] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0f3a6f96-1210-4da9-84c6-e8a137905680 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.990715] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4361b559-2dee-4f30-9f1e-339c9e6a8642 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.011913] env[63088]: WARNING nova.virt.vmwareapi.vmops [None req-02aa52fc-420f-46cd-9829-a68800c429e2 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e6a72b08-38b6-44ae-8911-6b39d50ee1c3 could not be found. [ 777.012132] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-02aa52fc-420f-46cd-9829-a68800c429e2 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 777.012311] env[63088]: INFO nova.compute.manager [None req-02aa52fc-420f-46cd-9829-a68800c429e2 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] Took 0.03 seconds to destroy the instance on the hypervisor. [ 777.012552] env[63088]: DEBUG oslo.service.loopingcall [None req-02aa52fc-420f-46cd-9829-a68800c429e2 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 777.012749] env[63088]: DEBUG nova.compute.manager [-] [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 777.012900] env[63088]: DEBUG nova.network.neutron [-] [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 777.026341] env[63088]: DEBUG nova.network.neutron [-] [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 777.251910] env[63088]: DEBUG nova.compute.manager [req-17f624d4-ea71-4249-87ef-27ee2a58dc4f req-ccbcfab6-be53-48e6-aa7a-3a344bddc7d2 service nova] [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] Received event network-changed-7876ef7c-e226-4a28-ac65-0d3647d7d12c {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 777.252153] env[63088]: DEBUG nova.compute.manager [req-17f624d4-ea71-4249-87ef-27ee2a58dc4f req-ccbcfab6-be53-48e6-aa7a-3a344bddc7d2 service nova] [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] Refreshing instance network info cache due to event network-changed-7876ef7c-e226-4a28-ac65-0d3647d7d12c. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 777.252401] env[63088]: DEBUG oslo_concurrency.lockutils [req-17f624d4-ea71-4249-87ef-27ee2a58dc4f req-ccbcfab6-be53-48e6-aa7a-3a344bddc7d2 service nova] Acquiring lock "refresh_cache-e6a72b08-38b6-44ae-8911-6b39d50ee1c3" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 777.252572] env[63088]: DEBUG oslo_concurrency.lockutils [req-17f624d4-ea71-4249-87ef-27ee2a58dc4f req-ccbcfab6-be53-48e6-aa7a-3a344bddc7d2 service nova] Acquired lock "refresh_cache-e6a72b08-38b6-44ae-8911-6b39d50ee1c3" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 777.252762] env[63088]: DEBUG nova.network.neutron [req-17f624d4-ea71-4249-87ef-27ee2a58dc4f req-ccbcfab6-be53-48e6-aa7a-3a344bddc7d2 service nova] [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] Refreshing network info cache for port 7876ef7c-e226-4a28-ac65-0d3647d7d12c {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 777.273574] env[63088]: DEBUG nova.network.neutron [None req-bce00a2e-7fb0-44a9-a66e-261b7c3d5384 tempest-ServerActionsV293TestJSON-1088890325 tempest-ServerActionsV293TestJSON-1088890325-project-member] [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 777.433488] env[63088]: DEBUG oslo_concurrency.lockutils [None req-1874938e-086f-4c3c-9049-55ea0246719c tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.802s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 777.434183] env[63088]: ERROR nova.compute.manager [None req-1874938e-086f-4c3c-9049-55ea0246719c tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 63967be2-f880-4637-9e38-8b244274427d, please check neutron logs for more information. [ 777.434183] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] Traceback (most recent call last): [ 777.434183] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 777.434183] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] self.driver.spawn(context, instance, image_meta, [ 777.434183] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 777.434183] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] self._vmops.spawn(context, instance, image_meta, injected_files, [ 777.434183] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 777.434183] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] vm_ref = self.build_virtual_machine(instance, [ 777.434183] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 777.434183] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] vif_infos = vmwarevif.get_vif_info(self._session, [ 777.434183] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 777.434568] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] for vif in network_info: [ 777.434568] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 777.434568] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] return self._sync_wrapper(fn, *args, **kwargs) [ 777.434568] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 777.434568] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] self.wait() [ 777.434568] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 777.434568] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] self[:] = self._gt.wait() [ 777.434568] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 777.434568] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] return self._exit_event.wait() [ 777.434568] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 777.434568] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] result = hub.switch() [ 777.434568] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 777.434568] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] return self.greenlet.switch() [ 777.434931] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 777.434931] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] result = function(*args, **kwargs) [ 777.434931] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 777.434931] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] return func(*args, **kwargs) [ 777.434931] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 777.434931] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] raise e [ 777.434931] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 777.434931] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] nwinfo = self.network_api.allocate_for_instance( [ 777.434931] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 777.434931] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] created_port_ids = self._update_ports_for_instance( [ 777.434931] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 777.434931] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] with excutils.save_and_reraise_exception(): [ 777.434931] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 777.435308] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] self.force_reraise() [ 777.435308] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 777.435308] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] raise self.value [ 777.435308] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 777.435308] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] updated_port = self._update_port( [ 777.435308] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 777.435308] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] _ensure_no_port_binding_failure(port) [ 777.435308] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 777.435308] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] raise exception.PortBindingFailed(port_id=port['id']) [ 777.435308] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] nova.exception.PortBindingFailed: Binding failed for port 63967be2-f880-4637-9e38-8b244274427d, please check neutron logs for more information. [ 777.435308] env[63088]: ERROR nova.compute.manager [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] [ 777.435613] env[63088]: DEBUG nova.compute.utils [None req-1874938e-086f-4c3c-9049-55ea0246719c tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] Binding failed for port 63967be2-f880-4637-9e38-8b244274427d, please check neutron logs for more information. {{(pid=63088) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 777.435970] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 15.916s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 777.436160] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 777.436310] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63088) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 777.436606] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3edaa40f-e8ca-4f6f-a9f2-5744f490f65d tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.651s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 777.439425] env[63088]: DEBUG nova.compute.manager [None req-1874938e-086f-4c3c-9049-55ea0246719c tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] Build of instance 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800 was re-scheduled: Binding failed for port 63967be2-f880-4637-9e38-8b244274427d, please check neutron logs for more information. {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 777.439854] env[63088]: DEBUG nova.compute.manager [None req-1874938e-086f-4c3c-9049-55ea0246719c tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] Unplugging VIFs for instance {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 777.440089] env[63088]: DEBUG oslo_concurrency.lockutils [None req-1874938e-086f-4c3c-9049-55ea0246719c tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] Acquiring lock "refresh_cache-1e1732ef-a198-46e7-8f5e-3b9d8fe8c800" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 777.440240] env[63088]: DEBUG oslo_concurrency.lockutils [None req-1874938e-086f-4c3c-9049-55ea0246719c tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] Acquired lock "refresh_cache-1e1732ef-a198-46e7-8f5e-3b9d8fe8c800" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 777.440392] env[63088]: DEBUG nova.network.neutron [None req-1874938e-086f-4c3c-9049-55ea0246719c tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 777.441747] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e343cd4-0a9e-473c-8d35-caa80c8ea036 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.450727] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d33d95b-2ff4-4db2-8df5-8d6ecd4bb9fd {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.465985] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b747969-b0e3-417f-b04d-6742147a133a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.472979] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-416b710f-d342-4357-a1c9-93eabd30ffcb {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.501741] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181477MB free_disk=140GB free_vcpus=48 pci_devices=None {{(pid=63088) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 777.502056] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 777.529198] env[63088]: DEBUG nova.network.neutron [-] [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 777.769574] env[63088]: DEBUG nova.network.neutron [req-17f624d4-ea71-4249-87ef-27ee2a58dc4f req-ccbcfab6-be53-48e6-aa7a-3a344bddc7d2 service nova] [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 777.775360] env[63088]: INFO nova.compute.manager [None req-bce00a2e-7fb0-44a9-a66e-261b7c3d5384 tempest-ServerActionsV293TestJSON-1088890325 tempest-ServerActionsV293TestJSON-1088890325-project-member] [instance: b82c89ad-f7d6-4c04-a48e-acdab29ea5e7] Took 1.02 seconds to deallocate network for instance. [ 777.820680] env[63088]: DEBUG nova.network.neutron [req-17f624d4-ea71-4249-87ef-27ee2a58dc4f req-ccbcfab6-be53-48e6-aa7a-3a344bddc7d2 service nova] [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 777.963708] env[63088]: DEBUG nova.network.neutron [None req-1874938e-086f-4c3c-9049-55ea0246719c tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 778.031472] env[63088]: INFO nova.compute.manager [-] [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] Took 1.02 seconds to deallocate network for instance. [ 778.033905] env[63088]: DEBUG nova.compute.claims [None req-02aa52fc-420f-46cd-9829-a68800c429e2 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] Aborting claim: {{(pid=63088) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 778.034127] env[63088]: DEBUG oslo_concurrency.lockutils [None req-02aa52fc-420f-46cd-9829-a68800c429e2 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 778.035978] env[63088]: DEBUG nova.network.neutron [None req-1874938e-086f-4c3c-9049-55ea0246719c tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 778.183732] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-780a803a-5e09-4e8c-ac43-3e9f9118bb36 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.191114] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c1ba84c-6a3a-46a9-9b57-2d3ca6dd2526 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.220652] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e269cce-5620-4d21-a33a-b3ec5a417c70 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.227524] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e181373-bfb5-4b42-9083-119a496ef7df {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.240156] env[63088]: DEBUG nova.compute.provider_tree [None req-3edaa40f-e8ca-4f6f-a9f2-5744f490f65d tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 778.324398] env[63088]: DEBUG oslo_concurrency.lockutils [req-17f624d4-ea71-4249-87ef-27ee2a58dc4f req-ccbcfab6-be53-48e6-aa7a-3a344bddc7d2 service nova] Releasing lock "refresh_cache-e6a72b08-38b6-44ae-8911-6b39d50ee1c3" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 778.324659] env[63088]: DEBUG nova.compute.manager [req-17f624d4-ea71-4249-87ef-27ee2a58dc4f req-ccbcfab6-be53-48e6-aa7a-3a344bddc7d2 service nova] [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] Received event network-vif-deleted-7876ef7c-e226-4a28-ac65-0d3647d7d12c {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 778.538027] env[63088]: DEBUG oslo_concurrency.lockutils [None req-1874938e-086f-4c3c-9049-55ea0246719c tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] Releasing lock "refresh_cache-1e1732ef-a198-46e7-8f5e-3b9d8fe8c800" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 778.538806] env[63088]: DEBUG nova.compute.manager [None req-1874938e-086f-4c3c-9049-55ea0246719c tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 778.538806] env[63088]: DEBUG nova.compute.manager [None req-1874938e-086f-4c3c-9049-55ea0246719c tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 778.538806] env[63088]: DEBUG nova.network.neutron [None req-1874938e-086f-4c3c-9049-55ea0246719c tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 778.553954] env[63088]: DEBUG nova.network.neutron [None req-1874938e-086f-4c3c-9049-55ea0246719c tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 778.742992] env[63088]: DEBUG nova.scheduler.client.report [None req-3edaa40f-e8ca-4f6f-a9f2-5744f490f65d tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 778.802634] env[63088]: INFO nova.scheduler.client.report [None req-bce00a2e-7fb0-44a9-a66e-261b7c3d5384 tempest-ServerActionsV293TestJSON-1088890325 tempest-ServerActionsV293TestJSON-1088890325-project-member] Deleted allocations for instance b82c89ad-f7d6-4c04-a48e-acdab29ea5e7 [ 779.057016] env[63088]: DEBUG nova.network.neutron [None req-1874938e-086f-4c3c-9049-55ea0246719c tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 779.247996] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3edaa40f-e8ca-4f6f-a9f2-5744f490f65d tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.811s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 779.248994] env[63088]: ERROR nova.compute.manager [None req-3edaa40f-e8ca-4f6f-a9f2-5744f490f65d tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 70b3718f-5593-4ace-be43-0e092ff9bfe7, please check neutron logs for more information. [ 779.248994] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] Traceback (most recent call last): [ 779.248994] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 779.248994] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] self.driver.spawn(context, instance, image_meta, [ 779.248994] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 779.248994] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 779.248994] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 779.248994] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] vm_ref = self.build_virtual_machine(instance, [ 779.248994] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 779.248994] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] vif_infos = vmwarevif.get_vif_info(self._session, [ 779.248994] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 779.249534] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] for vif in network_info: [ 779.249534] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 779.249534] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] return self._sync_wrapper(fn, *args, **kwargs) [ 779.249534] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 779.249534] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] self.wait() [ 779.249534] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 779.249534] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] self[:] = self._gt.wait() [ 779.249534] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 779.249534] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] return self._exit_event.wait() [ 779.249534] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 779.249534] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] result = hub.switch() [ 779.249534] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 779.249534] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] return self.greenlet.switch() [ 779.249974] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 779.249974] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] result = function(*args, **kwargs) [ 779.249974] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 779.249974] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] return func(*args, **kwargs) [ 779.249974] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 779.249974] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] raise e [ 779.249974] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 779.249974] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] nwinfo = self.network_api.allocate_for_instance( [ 779.249974] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 779.249974] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] created_port_ids = self._update_ports_for_instance( [ 779.249974] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 779.249974] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] with excutils.save_and_reraise_exception(): [ 779.249974] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 779.250813] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] self.force_reraise() [ 779.250813] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 779.250813] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] raise self.value [ 779.250813] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 779.250813] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] updated_port = self._update_port( [ 779.250813] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 779.250813] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] _ensure_no_port_binding_failure(port) [ 779.250813] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 779.250813] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] raise exception.PortBindingFailed(port_id=port['id']) [ 779.250813] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] nova.exception.PortBindingFailed: Binding failed for port 70b3718f-5593-4ace-be43-0e092ff9bfe7, please check neutron logs for more information. [ 779.250813] env[63088]: ERROR nova.compute.manager [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] [ 779.251204] env[63088]: DEBUG nova.compute.utils [None req-3edaa40f-e8ca-4f6f-a9f2-5744f490f65d tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] Binding failed for port 70b3718f-5593-4ace-be43-0e092ff9bfe7, please check neutron logs for more information. {{(pid=63088) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 779.251204] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3a58af89-5e90-4bc8-9f8f-c02b80d14311 tempest-InstanceActionsTestJSON-1717540784 tempest-InstanceActionsTestJSON-1717540784-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.219s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 779.253670] env[63088]: DEBUG nova.compute.manager [None req-3edaa40f-e8ca-4f6f-a9f2-5744f490f65d tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] Build of instance 17725a7b-ccd3-477e-a7a6-b96af45d6d1f was re-scheduled: Binding failed for port 70b3718f-5593-4ace-be43-0e092ff9bfe7, please check neutron logs for more information. {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 779.254104] env[63088]: DEBUG nova.compute.manager [None req-3edaa40f-e8ca-4f6f-a9f2-5744f490f65d tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] Unplugging VIFs for instance {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 779.254333] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3edaa40f-e8ca-4f6f-a9f2-5744f490f65d tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] Acquiring lock "refresh_cache-17725a7b-ccd3-477e-a7a6-b96af45d6d1f" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 779.254478] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3edaa40f-e8ca-4f6f-a9f2-5744f490f65d tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] Acquired lock "refresh_cache-17725a7b-ccd3-477e-a7a6-b96af45d6d1f" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.254637] env[63088]: DEBUG nova.network.neutron [None req-3edaa40f-e8ca-4f6f-a9f2-5744f490f65d tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 779.310576] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bce00a2e-7fb0-44a9-a66e-261b7c3d5384 tempest-ServerActionsV293TestJSON-1088890325 tempest-ServerActionsV293TestJSON-1088890325-project-member] Lock "b82c89ad-f7d6-4c04-a48e-acdab29ea5e7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 142.776s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 779.559861] env[63088]: INFO nova.compute.manager [None req-1874938e-086f-4c3c-9049-55ea0246719c tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] [instance: 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800] Took 1.02 seconds to deallocate network for instance. [ 779.775352] env[63088]: DEBUG nova.network.neutron [None req-3edaa40f-e8ca-4f6f-a9f2-5744f490f65d tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 779.815047] env[63088]: DEBUG nova.compute.manager [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: db032da3-d334-486f-a428-af1c8a3c360a] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 779.890724] env[63088]: DEBUG nova.network.neutron [None req-3edaa40f-e8ca-4f6f-a9f2-5744f490f65d tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 780.054096] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-361055e9-745a-4405-a109-b9b8a89bc234 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.061571] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-141004f1-dd1f-4de4-afef-17f2ac92be96 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.096459] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c3cb000-f8f3-4b13-8ca4-688e1f70f1ec {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.102330] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24eb6e68-610d-4290-b0bf-a132c9064bac {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.116022] env[63088]: DEBUG nova.compute.provider_tree [None req-3a58af89-5e90-4bc8-9f8f-c02b80d14311 tempest-InstanceActionsTestJSON-1717540784 tempest-InstanceActionsTestJSON-1717540784-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 780.337353] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 780.393638] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3edaa40f-e8ca-4f6f-a9f2-5744f490f65d tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] Releasing lock "refresh_cache-17725a7b-ccd3-477e-a7a6-b96af45d6d1f" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 780.393881] env[63088]: DEBUG nova.compute.manager [None req-3edaa40f-e8ca-4f6f-a9f2-5744f490f65d tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 780.394077] env[63088]: DEBUG nova.compute.manager [None req-3edaa40f-e8ca-4f6f-a9f2-5744f490f65d tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 780.394250] env[63088]: DEBUG nova.network.neutron [None req-3edaa40f-e8ca-4f6f-a9f2-5744f490f65d tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 780.409125] env[63088]: DEBUG nova.network.neutron [None req-3edaa40f-e8ca-4f6f-a9f2-5744f490f65d tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 780.600768] env[63088]: INFO nova.scheduler.client.report [None req-1874938e-086f-4c3c-9049-55ea0246719c tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] Deleted allocations for instance 1e1732ef-a198-46e7-8f5e-3b9d8fe8c800 [ 780.624074] env[63088]: DEBUG nova.scheduler.client.report [None req-3a58af89-5e90-4bc8-9f8f-c02b80d14311 tempest-InstanceActionsTestJSON-1717540784 tempest-InstanceActionsTestJSON-1717540784-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 780.911707] env[63088]: DEBUG nova.network.neutron [None req-3edaa40f-e8ca-4f6f-a9f2-5744f490f65d tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 781.113698] env[63088]: DEBUG oslo_concurrency.lockutils [None req-1874938e-086f-4c3c-9049-55ea0246719c tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] Lock "1e1732ef-a198-46e7-8f5e-3b9d8fe8c800" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 135.919s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 781.130965] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3a58af89-5e90-4bc8-9f8f-c02b80d14311 tempest-InstanceActionsTestJSON-1717540784 tempest-InstanceActionsTestJSON-1717540784-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.880s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 781.133372] env[63088]: ERROR nova.compute.manager [None req-3a58af89-5e90-4bc8-9f8f-c02b80d14311 tempest-InstanceActionsTestJSON-1717540784 tempest-InstanceActionsTestJSON-1717540784-project-member] [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6904053f-859c-4b6f-80fc-0278d414bf42, please check neutron logs for more information. [ 781.133372] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] Traceback (most recent call last): [ 781.133372] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 781.133372] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] self.driver.spawn(context, instance, image_meta, [ 781.133372] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 781.133372] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] self._vmops.spawn(context, instance, image_meta, injected_files, [ 781.133372] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 781.133372] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] vm_ref = self.build_virtual_machine(instance, [ 781.133372] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 781.133372] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] vif_infos = vmwarevif.get_vif_info(self._session, [ 781.133372] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 781.133766] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] for vif in network_info: [ 781.133766] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 781.133766] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] return self._sync_wrapper(fn, *args, **kwargs) [ 781.133766] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 781.133766] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] self.wait() [ 781.133766] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 781.133766] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] self[:] = self._gt.wait() [ 781.133766] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 781.133766] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] return self._exit_event.wait() [ 781.133766] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 781.133766] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] result = hub.switch() [ 781.133766] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 781.133766] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] return self.greenlet.switch() [ 781.134146] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 781.134146] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] result = function(*args, **kwargs) [ 781.134146] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 781.134146] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] return func(*args, **kwargs) [ 781.134146] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 781.134146] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] raise e [ 781.134146] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 781.134146] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] nwinfo = self.network_api.allocate_for_instance( [ 781.134146] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 781.134146] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] created_port_ids = self._update_ports_for_instance( [ 781.134146] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 781.134146] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] with excutils.save_and_reraise_exception(): [ 781.134146] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 781.134509] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] self.force_reraise() [ 781.134509] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 781.134509] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] raise self.value [ 781.134509] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 781.134509] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] updated_port = self._update_port( [ 781.134509] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 781.134509] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] _ensure_no_port_binding_failure(port) [ 781.134509] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 781.134509] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] raise exception.PortBindingFailed(port_id=port['id']) [ 781.134509] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] nova.exception.PortBindingFailed: Binding failed for port 6904053f-859c-4b6f-80fc-0278d414bf42, please check neutron logs for more information. [ 781.134509] env[63088]: ERROR nova.compute.manager [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] [ 781.134905] env[63088]: DEBUG nova.compute.utils [None req-3a58af89-5e90-4bc8-9f8f-c02b80d14311 tempest-InstanceActionsTestJSON-1717540784 tempest-InstanceActionsTestJSON-1717540784-project-member] [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] Binding failed for port 6904053f-859c-4b6f-80fc-0278d414bf42, please check neutron logs for more information. {{(pid=63088) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 781.135812] env[63088]: DEBUG nova.compute.manager [None req-3a58af89-5e90-4bc8-9f8f-c02b80d14311 tempest-InstanceActionsTestJSON-1717540784 tempest-InstanceActionsTestJSON-1717540784-project-member] [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] Build of instance 7b5cc7b5-98b9-4344-b714-270dca798799 was re-scheduled: Binding failed for port 6904053f-859c-4b6f-80fc-0278d414bf42, please check neutron logs for more information. {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 781.135928] env[63088]: DEBUG nova.compute.manager [None req-3a58af89-5e90-4bc8-9f8f-c02b80d14311 tempest-InstanceActionsTestJSON-1717540784 tempest-InstanceActionsTestJSON-1717540784-project-member] [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] Unplugging VIFs for instance {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 781.136239] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3a58af89-5e90-4bc8-9f8f-c02b80d14311 tempest-InstanceActionsTestJSON-1717540784 tempest-InstanceActionsTestJSON-1717540784-project-member] Acquiring lock "refresh_cache-7b5cc7b5-98b9-4344-b714-270dca798799" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 781.136322] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3a58af89-5e90-4bc8-9f8f-c02b80d14311 tempest-InstanceActionsTestJSON-1717540784 tempest-InstanceActionsTestJSON-1717540784-project-member] Acquired lock "refresh_cache-7b5cc7b5-98b9-4344-b714-270dca798799" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 781.137044] env[63088]: DEBUG nova.network.neutron [None req-3a58af89-5e90-4bc8-9f8f-c02b80d14311 tempest-InstanceActionsTestJSON-1717540784 tempest-InstanceActionsTestJSON-1717540784-project-member] [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 781.138432] env[63088]: DEBUG oslo_concurrency.lockutils [None req-87ab5d07-9ebb-429d-8640-0197960e7d1a tempest-ServersNegativeTestMultiTenantJSON-2054426678 tempest-ServersNegativeTestMultiTenantJSON-2054426678-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.619s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 781.414542] env[63088]: INFO nova.compute.manager [None req-3edaa40f-e8ca-4f6f-a9f2-5744f490f65d tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] [instance: 17725a7b-ccd3-477e-a7a6-b96af45d6d1f] Took 1.02 seconds to deallocate network for instance. [ 781.617603] env[63088]: DEBUG nova.compute.manager [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 11ec987f-9908-4366-81ae-cbfe4de67b2d] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 781.656423] env[63088]: DEBUG nova.network.neutron [None req-3a58af89-5e90-4bc8-9f8f-c02b80d14311 tempest-InstanceActionsTestJSON-1717540784 tempest-InstanceActionsTestJSON-1717540784-project-member] [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 781.753623] env[63088]: DEBUG nova.network.neutron [None req-3a58af89-5e90-4bc8-9f8f-c02b80d14311 tempest-InstanceActionsTestJSON-1717540784 tempest-InstanceActionsTestJSON-1717540784-project-member] [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 781.872206] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa940d91-6d09-43c7-8730-333b5cd3159a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.880097] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4a13fdc-772b-4a14-ae80-baffb13213b1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.917084] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ba5d805-8e3c-489d-ad8d-84e9368f4532 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.927545] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-893ed2f9-da87-4000-b320-f467c1326301 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.942150] env[63088]: DEBUG nova.compute.provider_tree [None req-87ab5d07-9ebb-429d-8640-0197960e7d1a tempest-ServersNegativeTestMultiTenantJSON-2054426678 tempest-ServersNegativeTestMultiTenantJSON-2054426678-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 782.144304] env[63088]: DEBUG oslo_concurrency.lockutils [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 782.255948] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3a58af89-5e90-4bc8-9f8f-c02b80d14311 tempest-InstanceActionsTestJSON-1717540784 tempest-InstanceActionsTestJSON-1717540784-project-member] Releasing lock "refresh_cache-7b5cc7b5-98b9-4344-b714-270dca798799" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 782.256212] env[63088]: DEBUG nova.compute.manager [None req-3a58af89-5e90-4bc8-9f8f-c02b80d14311 tempest-InstanceActionsTestJSON-1717540784 tempest-InstanceActionsTestJSON-1717540784-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 782.256399] env[63088]: DEBUG nova.compute.manager [None req-3a58af89-5e90-4bc8-9f8f-c02b80d14311 tempest-InstanceActionsTestJSON-1717540784 tempest-InstanceActionsTestJSON-1717540784-project-member] [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 782.257180] env[63088]: DEBUG nova.network.neutron [None req-3a58af89-5e90-4bc8-9f8f-c02b80d14311 tempest-InstanceActionsTestJSON-1717540784 tempest-InstanceActionsTestJSON-1717540784-project-member] [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 782.274397] env[63088]: DEBUG nova.network.neutron [None req-3a58af89-5e90-4bc8-9f8f-c02b80d14311 tempest-InstanceActionsTestJSON-1717540784 tempest-InstanceActionsTestJSON-1717540784-project-member] [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 782.447016] env[63088]: INFO nova.scheduler.client.report [None req-3edaa40f-e8ca-4f6f-a9f2-5744f490f65d tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] Deleted allocations for instance 17725a7b-ccd3-477e-a7a6-b96af45d6d1f [ 782.450978] env[63088]: DEBUG nova.scheduler.client.report [None req-87ab5d07-9ebb-429d-8640-0197960e7d1a tempest-ServersNegativeTestMultiTenantJSON-2054426678 tempest-ServersNegativeTestMultiTenantJSON-2054426678-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 782.779028] env[63088]: DEBUG nova.network.neutron [None req-3a58af89-5e90-4bc8-9f8f-c02b80d14311 tempest-InstanceActionsTestJSON-1717540784 tempest-InstanceActionsTestJSON-1717540784-project-member] [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 782.956530] env[63088]: DEBUG oslo_concurrency.lockutils [None req-87ab5d07-9ebb-429d-8640-0197960e7d1a tempest-ServersNegativeTestMultiTenantJSON-2054426678 tempest-ServersNegativeTestMultiTenantJSON-2054426678-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.818s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 782.957462] env[63088]: ERROR nova.compute.manager [None req-87ab5d07-9ebb-429d-8640-0197960e7d1a tempest-ServersNegativeTestMultiTenantJSON-2054426678 tempest-ServersNegativeTestMultiTenantJSON-2054426678-project-member] [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5c2db8b3-531e-4a33-a138-03a6256495c1, please check neutron logs for more information. [ 782.957462] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] Traceback (most recent call last): [ 782.957462] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 782.957462] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] self.driver.spawn(context, instance, image_meta, [ 782.957462] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 782.957462] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] self._vmops.spawn(context, instance, image_meta, injected_files, [ 782.957462] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 782.957462] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] vm_ref = self.build_virtual_machine(instance, [ 782.957462] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 782.957462] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] vif_infos = vmwarevif.get_vif_info(self._session, [ 782.957462] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 782.957859] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] for vif in network_info: [ 782.957859] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 782.957859] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] return self._sync_wrapper(fn, *args, **kwargs) [ 782.957859] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 782.957859] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] self.wait() [ 782.957859] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 782.957859] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] self[:] = self._gt.wait() [ 782.957859] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 782.957859] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] return self._exit_event.wait() [ 782.957859] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 782.957859] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] result = hub.switch() [ 782.957859] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 782.957859] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] return self.greenlet.switch() [ 782.958315] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 782.958315] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] result = function(*args, **kwargs) [ 782.958315] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 782.958315] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] return func(*args, **kwargs) [ 782.958315] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 782.958315] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] raise e [ 782.958315] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 782.958315] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] nwinfo = self.network_api.allocate_for_instance( [ 782.958315] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 782.958315] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] created_port_ids = self._update_ports_for_instance( [ 782.958315] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 782.958315] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] with excutils.save_and_reraise_exception(): [ 782.958315] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 782.958745] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] self.force_reraise() [ 782.958745] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 782.958745] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] raise self.value [ 782.958745] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 782.958745] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] updated_port = self._update_port( [ 782.958745] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 782.958745] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] _ensure_no_port_binding_failure(port) [ 782.958745] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 782.958745] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] raise exception.PortBindingFailed(port_id=port['id']) [ 782.958745] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] nova.exception.PortBindingFailed: Binding failed for port 5c2db8b3-531e-4a33-a138-03a6256495c1, please check neutron logs for more information. [ 782.958745] env[63088]: ERROR nova.compute.manager [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] [ 782.959120] env[63088]: DEBUG nova.compute.utils [None req-87ab5d07-9ebb-429d-8640-0197960e7d1a tempest-ServersNegativeTestMultiTenantJSON-2054426678 tempest-ServersNegativeTestMultiTenantJSON-2054426678-project-member] [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] Binding failed for port 5c2db8b3-531e-4a33-a138-03a6256495c1, please check neutron logs for more information. {{(pid=63088) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 782.960812] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3edaa40f-e8ca-4f6f-a9f2-5744f490f65d tempest-ServerRescueNegativeTestJSON-251937651 tempest-ServerRescueNegativeTestJSON-251937651-project-member] Lock "17725a7b-ccd3-477e-a7a6-b96af45d6d1f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 136.930s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 782.961324] env[63088]: DEBUG nova.compute.manager [None req-87ab5d07-9ebb-429d-8640-0197960e7d1a tempest-ServersNegativeTestMultiTenantJSON-2054426678 tempest-ServersNegativeTestMultiTenantJSON-2054426678-project-member] [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] Build of instance d851a26b-48aa-4ffd-9c23-cea27543e0fa was re-scheduled: Binding failed for port 5c2db8b3-531e-4a33-a138-03a6256495c1, please check neutron logs for more information. {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 782.961775] env[63088]: DEBUG nova.compute.manager [None req-87ab5d07-9ebb-429d-8640-0197960e7d1a tempest-ServersNegativeTestMultiTenantJSON-2054426678 tempest-ServersNegativeTestMultiTenantJSON-2054426678-project-member] [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] Unplugging VIFs for instance {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 782.962010] env[63088]: DEBUG oslo_concurrency.lockutils [None req-87ab5d07-9ebb-429d-8640-0197960e7d1a tempest-ServersNegativeTestMultiTenantJSON-2054426678 tempest-ServersNegativeTestMultiTenantJSON-2054426678-project-member] Acquiring lock "refresh_cache-d851a26b-48aa-4ffd-9c23-cea27543e0fa" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 782.962169] env[63088]: DEBUG oslo_concurrency.lockutils [None req-87ab5d07-9ebb-429d-8640-0197960e7d1a tempest-ServersNegativeTestMultiTenantJSON-2054426678 tempest-ServersNegativeTestMultiTenantJSON-2054426678-project-member] Acquired lock "refresh_cache-d851a26b-48aa-4ffd-9c23-cea27543e0fa" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.962318] env[63088]: DEBUG nova.network.neutron [None req-87ab5d07-9ebb-429d-8640-0197960e7d1a tempest-ServersNegativeTestMultiTenantJSON-2054426678 tempest-ServersNegativeTestMultiTenantJSON-2054426678-project-member] [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 782.963435] env[63088]: DEBUG oslo_concurrency.lockutils [None req-aa39901a-ab16-4a54-a2f6-147052a0ed90 tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.938s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 783.282277] env[63088]: INFO nova.compute.manager [None req-3a58af89-5e90-4bc8-9f8f-c02b80d14311 tempest-InstanceActionsTestJSON-1717540784 tempest-InstanceActionsTestJSON-1717540784-project-member] [instance: 7b5cc7b5-98b9-4344-b714-270dca798799] Took 1.03 seconds to deallocate network for instance. [ 783.472999] env[63088]: DEBUG nova.compute.manager [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 783.496201] env[63088]: DEBUG nova.network.neutron [None req-87ab5d07-9ebb-429d-8640-0197960e7d1a tempest-ServersNegativeTestMultiTenantJSON-2054426678 tempest-ServersNegativeTestMultiTenantJSON-2054426678-project-member] [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 783.687764] env[63088]: DEBUG nova.network.neutron [None req-87ab5d07-9ebb-429d-8640-0197960e7d1a tempest-ServersNegativeTestMultiTenantJSON-2054426678 tempest-ServersNegativeTestMultiTenantJSON-2054426678-project-member] [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 783.718183] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28da35e6-defc-4a91-82da-0276b5546ee6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.726741] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b080571-06f4-4320-ac1b-9c8af2df6793 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.757701] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc779ce7-60a3-444a-a81c-2e4f42dd560a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.765622] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2db7f75a-5286-4424-842d-88a5b7237632 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.779162] env[63088]: DEBUG nova.compute.provider_tree [None req-aa39901a-ab16-4a54-a2f6-147052a0ed90 tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 783.996649] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 784.190634] env[63088]: DEBUG oslo_concurrency.lockutils [None req-87ab5d07-9ebb-429d-8640-0197960e7d1a tempest-ServersNegativeTestMultiTenantJSON-2054426678 tempest-ServersNegativeTestMultiTenantJSON-2054426678-project-member] Releasing lock "refresh_cache-d851a26b-48aa-4ffd-9c23-cea27543e0fa" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 784.190910] env[63088]: DEBUG nova.compute.manager [None req-87ab5d07-9ebb-429d-8640-0197960e7d1a tempest-ServersNegativeTestMultiTenantJSON-2054426678 tempest-ServersNegativeTestMultiTenantJSON-2054426678-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 784.191111] env[63088]: DEBUG nova.compute.manager [None req-87ab5d07-9ebb-429d-8640-0197960e7d1a tempest-ServersNegativeTestMultiTenantJSON-2054426678 tempest-ServersNegativeTestMultiTenantJSON-2054426678-project-member] [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 784.191284] env[63088]: DEBUG nova.network.neutron [None req-87ab5d07-9ebb-429d-8640-0197960e7d1a tempest-ServersNegativeTestMultiTenantJSON-2054426678 tempest-ServersNegativeTestMultiTenantJSON-2054426678-project-member] [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 784.219225] env[63088]: DEBUG nova.network.neutron [None req-87ab5d07-9ebb-429d-8640-0197960e7d1a tempest-ServersNegativeTestMultiTenantJSON-2054426678 tempest-ServersNegativeTestMultiTenantJSON-2054426678-project-member] [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 784.282542] env[63088]: DEBUG nova.scheduler.client.report [None req-aa39901a-ab16-4a54-a2f6-147052a0ed90 tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 784.316241] env[63088]: INFO nova.scheduler.client.report [None req-3a58af89-5e90-4bc8-9f8f-c02b80d14311 tempest-InstanceActionsTestJSON-1717540784 tempest-InstanceActionsTestJSON-1717540784-project-member] Deleted allocations for instance 7b5cc7b5-98b9-4344-b714-270dca798799 [ 784.725601] env[63088]: DEBUG nova.network.neutron [None req-87ab5d07-9ebb-429d-8640-0197960e7d1a tempest-ServersNegativeTestMultiTenantJSON-2054426678 tempest-ServersNegativeTestMultiTenantJSON-2054426678-project-member] [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 784.788981] env[63088]: DEBUG oslo_concurrency.lockutils [None req-aa39901a-ab16-4a54-a2f6-147052a0ed90 tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.824s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 784.788981] env[63088]: ERROR nova.compute.manager [None req-aa39901a-ab16-4a54-a2f6-147052a0ed90 tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] [instance: 43ded929-6890-4296-ae90-230f7c8b2595] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a0d4f41f-99bf-4a7a-9441-0835bc236922, please check neutron logs for more information. [ 784.788981] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] Traceback (most recent call last): [ 784.788981] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 784.788981] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] self.driver.spawn(context, instance, image_meta, [ 784.788981] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 784.788981] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] self._vmops.spawn(context, instance, image_meta, injected_files, [ 784.788981] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 784.788981] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] vm_ref = self.build_virtual_machine(instance, [ 784.789622] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 784.789622] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] vif_infos = vmwarevif.get_vif_info(self._session, [ 784.789622] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 784.789622] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] for vif in network_info: [ 784.789622] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 784.789622] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] return self._sync_wrapper(fn, *args, **kwargs) [ 784.789622] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 784.789622] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] self.wait() [ 784.789622] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 784.789622] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] self[:] = self._gt.wait() [ 784.789622] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 784.789622] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] return self._exit_event.wait() [ 784.789622] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 784.790281] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] current.throw(*self._exc) [ 784.790281] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 784.790281] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] result = function(*args, **kwargs) [ 784.790281] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 784.790281] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] return func(*args, **kwargs) [ 784.790281] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 784.790281] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] raise e [ 784.790281] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 784.790281] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] nwinfo = self.network_api.allocate_for_instance( [ 784.790281] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 784.790281] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] created_port_ids = self._update_ports_for_instance( [ 784.790281] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 784.790281] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] with excutils.save_and_reraise_exception(): [ 784.790695] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 784.790695] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] self.force_reraise() [ 784.790695] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 784.790695] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] raise self.value [ 784.790695] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 784.790695] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] updated_port = self._update_port( [ 784.790695] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 784.790695] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] _ensure_no_port_binding_failure(port) [ 784.790695] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 784.790695] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] raise exception.PortBindingFailed(port_id=port['id']) [ 784.790695] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] nova.exception.PortBindingFailed: Binding failed for port a0d4f41f-99bf-4a7a-9441-0835bc236922, please check neutron logs for more information. [ 784.790695] env[63088]: ERROR nova.compute.manager [instance: 43ded929-6890-4296-ae90-230f7c8b2595] [ 784.791156] env[63088]: DEBUG nova.compute.utils [None req-aa39901a-ab16-4a54-a2f6-147052a0ed90 tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] [instance: 43ded929-6890-4296-ae90-230f7c8b2595] Binding failed for port a0d4f41f-99bf-4a7a-9441-0835bc236922, please check neutron logs for more information. {{(pid=63088) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 784.791156] env[63088]: DEBUG oslo_concurrency.lockutils [None req-1c0c7b6c-8782-4a52-b8ee-b7fd4463c345 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.232s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 784.795562] env[63088]: DEBUG nova.compute.manager [None req-aa39901a-ab16-4a54-a2f6-147052a0ed90 tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] [instance: 43ded929-6890-4296-ae90-230f7c8b2595] Build of instance 43ded929-6890-4296-ae90-230f7c8b2595 was re-scheduled: Binding failed for port a0d4f41f-99bf-4a7a-9441-0835bc236922, please check neutron logs for more information. {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 784.795562] env[63088]: DEBUG nova.compute.manager [None req-aa39901a-ab16-4a54-a2f6-147052a0ed90 tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] [instance: 43ded929-6890-4296-ae90-230f7c8b2595] Unplugging VIFs for instance {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 784.795562] env[63088]: DEBUG oslo_concurrency.lockutils [None req-aa39901a-ab16-4a54-a2f6-147052a0ed90 tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] Acquiring lock "refresh_cache-43ded929-6890-4296-ae90-230f7c8b2595" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 784.795562] env[63088]: DEBUG oslo_concurrency.lockutils [None req-aa39901a-ab16-4a54-a2f6-147052a0ed90 tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] Acquired lock "refresh_cache-43ded929-6890-4296-ae90-230f7c8b2595" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 784.795821] env[63088]: DEBUG nova.network.neutron [None req-aa39901a-ab16-4a54-a2f6-147052a0ed90 tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] [instance: 43ded929-6890-4296-ae90-230f7c8b2595] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 784.827754] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3a58af89-5e90-4bc8-9f8f-c02b80d14311 tempest-InstanceActionsTestJSON-1717540784 tempest-InstanceActionsTestJSON-1717540784-project-member] Lock "7b5cc7b5-98b9-4344-b714-270dca798799" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 137.097s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 785.230431] env[63088]: INFO nova.compute.manager [None req-87ab5d07-9ebb-429d-8640-0197960e7d1a tempest-ServersNegativeTestMultiTenantJSON-2054426678 tempest-ServersNegativeTestMultiTenantJSON-2054426678-project-member] [instance: d851a26b-48aa-4ffd-9c23-cea27543e0fa] Took 1.04 seconds to deallocate network for instance. [ 785.336190] env[63088]: DEBUG nova.compute.manager [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 785.375058] env[63088]: DEBUG nova.network.neutron [None req-aa39901a-ab16-4a54-a2f6-147052a0ed90 tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] [instance: 43ded929-6890-4296-ae90-230f7c8b2595] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 785.488728] env[63088]: DEBUG nova.network.neutron [None req-aa39901a-ab16-4a54-a2f6-147052a0ed90 tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] [instance: 43ded929-6890-4296-ae90-230f7c8b2595] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 785.571033] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6a090ca-550f-4c92-8a12-3cd9df05b999 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.580016] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d566850a-1685-443f-aba0-3d64128e2a6d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.610317] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee2e725f-e410-425b-9f5f-7f6ef545a718 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.617505] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bad9ea7f-2ac0-4de4-8572-8d05f0c93f80 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.631897] env[63088]: DEBUG nova.compute.provider_tree [None req-1c0c7b6c-8782-4a52-b8ee-b7fd4463c345 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 785.804747] env[63088]: DEBUG oslo_concurrency.lockutils [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Acquiring lock "2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 785.805362] env[63088]: DEBUG oslo_concurrency.lockutils [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Lock "2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 785.864409] env[63088]: DEBUG oslo_concurrency.lockutils [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 785.991909] env[63088]: DEBUG oslo_concurrency.lockutils [None req-aa39901a-ab16-4a54-a2f6-147052a0ed90 tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] Releasing lock "refresh_cache-43ded929-6890-4296-ae90-230f7c8b2595" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 785.993471] env[63088]: DEBUG nova.compute.manager [None req-aa39901a-ab16-4a54-a2f6-147052a0ed90 tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 785.993722] env[63088]: DEBUG nova.compute.manager [None req-aa39901a-ab16-4a54-a2f6-147052a0ed90 tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] [instance: 43ded929-6890-4296-ae90-230f7c8b2595] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 785.993900] env[63088]: DEBUG nova.network.neutron [None req-aa39901a-ab16-4a54-a2f6-147052a0ed90 tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] [instance: 43ded929-6890-4296-ae90-230f7c8b2595] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 786.015478] env[63088]: DEBUG nova.network.neutron [None req-aa39901a-ab16-4a54-a2f6-147052a0ed90 tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] [instance: 43ded929-6890-4296-ae90-230f7c8b2595] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 786.135070] env[63088]: DEBUG nova.scheduler.client.report [None req-1c0c7b6c-8782-4a52-b8ee-b7fd4463c345 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 786.277021] env[63088]: INFO nova.scheduler.client.report [None req-87ab5d07-9ebb-429d-8640-0197960e7d1a tempest-ServersNegativeTestMultiTenantJSON-2054426678 tempest-ServersNegativeTestMultiTenantJSON-2054426678-project-member] Deleted allocations for instance d851a26b-48aa-4ffd-9c23-cea27543e0fa [ 786.523424] env[63088]: DEBUG nova.network.neutron [None req-aa39901a-ab16-4a54-a2f6-147052a0ed90 tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] [instance: 43ded929-6890-4296-ae90-230f7c8b2595] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 786.642221] env[63088]: DEBUG oslo_concurrency.lockutils [None req-1c0c7b6c-8782-4a52-b8ee-b7fd4463c345 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.850s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 786.642221] env[63088]: ERROR nova.compute.manager [None req-1c0c7b6c-8782-4a52-b8ee-b7fd4463c345 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 09226e11-0d58-4ff6-adfe-80141e1092c2, please check neutron logs for more information. [ 786.642221] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] Traceback (most recent call last): [ 786.642221] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 786.642221] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] self.driver.spawn(context, instance, image_meta, [ 786.642221] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 786.642221] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 786.642221] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 786.642221] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] vm_ref = self.build_virtual_machine(instance, [ 786.642654] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 786.642654] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] vif_infos = vmwarevif.get_vif_info(self._session, [ 786.642654] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 786.642654] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] for vif in network_info: [ 786.642654] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 786.642654] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] return self._sync_wrapper(fn, *args, **kwargs) [ 786.642654] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 786.642654] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] self.wait() [ 786.642654] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 786.642654] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] self[:] = self._gt.wait() [ 786.642654] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 786.642654] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] return self._exit_event.wait() [ 786.642654] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 786.643067] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] current.throw(*self._exc) [ 786.643067] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 786.643067] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] result = function(*args, **kwargs) [ 786.643067] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 786.643067] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] return func(*args, **kwargs) [ 786.643067] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 786.643067] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] raise e [ 786.643067] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 786.643067] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] nwinfo = self.network_api.allocate_for_instance( [ 786.643067] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 786.643067] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] created_port_ids = self._update_ports_for_instance( [ 786.643067] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 786.643067] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] with excutils.save_and_reraise_exception(): [ 786.643444] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 786.643444] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] self.force_reraise() [ 786.643444] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 786.643444] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] raise self.value [ 786.643444] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 786.643444] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] updated_port = self._update_port( [ 786.643444] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 786.643444] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] _ensure_no_port_binding_failure(port) [ 786.643444] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 786.643444] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] raise exception.PortBindingFailed(port_id=port['id']) [ 786.643444] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] nova.exception.PortBindingFailed: Binding failed for port 09226e11-0d58-4ff6-adfe-80141e1092c2, please check neutron logs for more information. [ 786.643444] env[63088]: ERROR nova.compute.manager [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] [ 786.643792] env[63088]: DEBUG nova.compute.utils [None req-1c0c7b6c-8782-4a52-b8ee-b7fd4463c345 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] Binding failed for port 09226e11-0d58-4ff6-adfe-80141e1092c2, please check neutron logs for more information. {{(pid=63088) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 786.643865] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b5c531a2-a229-4998-9a9b-9ebb1e7afff2 tempest-ServerGroupTestJSON-1534054659 tempest-ServerGroupTestJSON-1534054659-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.795s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 786.646235] env[63088]: INFO nova.compute.claims [None req-b5c531a2-a229-4998-9a9b-9ebb1e7afff2 tempest-ServerGroupTestJSON-1534054659 tempest-ServerGroupTestJSON-1534054659-project-member] [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 786.648782] env[63088]: DEBUG nova.compute.manager [None req-1c0c7b6c-8782-4a52-b8ee-b7fd4463c345 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] Build of instance fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3 was re-scheduled: Binding failed for port 09226e11-0d58-4ff6-adfe-80141e1092c2, please check neutron logs for more information. {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 786.649250] env[63088]: DEBUG nova.compute.manager [None req-1c0c7b6c-8782-4a52-b8ee-b7fd4463c345 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] Unplugging VIFs for instance {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 786.649494] env[63088]: DEBUG oslo_concurrency.lockutils [None req-1c0c7b6c-8782-4a52-b8ee-b7fd4463c345 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Acquiring lock "refresh_cache-fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 786.650608] env[63088]: DEBUG oslo_concurrency.lockutils [None req-1c0c7b6c-8782-4a52-b8ee-b7fd4463c345 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Acquired lock "refresh_cache-fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 786.650608] env[63088]: DEBUG nova.network.neutron [None req-1c0c7b6c-8782-4a52-b8ee-b7fd4463c345 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 786.785290] env[63088]: DEBUG oslo_concurrency.lockutils [None req-87ab5d07-9ebb-429d-8640-0197960e7d1a tempest-ServersNegativeTestMultiTenantJSON-2054426678 tempest-ServersNegativeTestMultiTenantJSON-2054426678-project-member] Lock "d851a26b-48aa-4ffd-9c23-cea27543e0fa" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 136.712s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 787.025554] env[63088]: INFO nova.compute.manager [None req-aa39901a-ab16-4a54-a2f6-147052a0ed90 tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] [instance: 43ded929-6890-4296-ae90-230f7c8b2595] Took 1.03 seconds to deallocate network for instance. [ 787.186120] env[63088]: DEBUG nova.network.neutron [None req-1c0c7b6c-8782-4a52-b8ee-b7fd4463c345 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 787.288170] env[63088]: DEBUG nova.compute.manager [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: e845f60b-004f-4a5c-813e-0c183e99eba6] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 787.326090] env[63088]: DEBUG nova.network.neutron [None req-1c0c7b6c-8782-4a52-b8ee-b7fd4463c345 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 787.807885] env[63088]: DEBUG oslo_concurrency.lockutils [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 787.829950] env[63088]: DEBUG oslo_concurrency.lockutils [None req-1c0c7b6c-8782-4a52-b8ee-b7fd4463c345 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Releasing lock "refresh_cache-fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 787.829950] env[63088]: DEBUG nova.compute.manager [None req-1c0c7b6c-8782-4a52-b8ee-b7fd4463c345 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 787.829950] env[63088]: DEBUG nova.compute.manager [None req-1c0c7b6c-8782-4a52-b8ee-b7fd4463c345 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 787.830560] env[63088]: DEBUG nova.network.neutron [None req-1c0c7b6c-8782-4a52-b8ee-b7fd4463c345 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 787.874780] env[63088]: DEBUG nova.network.neutron [None req-1c0c7b6c-8782-4a52-b8ee-b7fd4463c345 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 787.902095] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8d208c7-93ad-4e7c-b97b-61579d4ff5e5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.910143] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e35765dc-d5e0-4c0e-b450-5fc7a33d8a70 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.942747] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-820d1518-d95f-4f38-b394-634489815735 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.950975] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd03e305-718b-41fe-b966-012ffa9181e0 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.964865] env[63088]: DEBUG nova.compute.provider_tree [None req-b5c531a2-a229-4998-9a9b-9ebb1e7afff2 tempest-ServerGroupTestJSON-1534054659 tempest-ServerGroupTestJSON-1534054659-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 788.054430] env[63088]: INFO nova.scheduler.client.report [None req-aa39901a-ab16-4a54-a2f6-147052a0ed90 tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] Deleted allocations for instance 43ded929-6890-4296-ae90-230f7c8b2595 [ 788.378608] env[63088]: DEBUG nova.network.neutron [None req-1c0c7b6c-8782-4a52-b8ee-b7fd4463c345 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 788.468339] env[63088]: DEBUG nova.scheduler.client.report [None req-b5c531a2-a229-4998-9a9b-9ebb1e7afff2 tempest-ServerGroupTestJSON-1534054659 tempest-ServerGroupTestJSON-1534054659-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 788.567646] env[63088]: DEBUG oslo_concurrency.lockutils [None req-aa39901a-ab16-4a54-a2f6-147052a0ed90 tempest-SecurityGroupsTestJSON-1921597564 tempest-SecurityGroupsTestJSON-1921597564-project-member] Lock "43ded929-6890-4296-ae90-230f7c8b2595" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 136.225s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 788.883266] env[63088]: INFO nova.compute.manager [None req-1c0c7b6c-8782-4a52-b8ee-b7fd4463c345 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3] Took 1.05 seconds to deallocate network for instance. [ 788.973540] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b5c531a2-a229-4998-9a9b-9ebb1e7afff2 tempest-ServerGroupTestJSON-1534054659 tempest-ServerGroupTestJSON-1534054659-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.330s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 788.974778] env[63088]: DEBUG oslo_concurrency.lockutils [None req-96c00a75-0383-4506-a4c6-d540aa18aaef tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.493s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 789.071027] env[63088]: DEBUG nova.compute.manager [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] [instance: 4f6e2d99-7300-42e6-8abf-362ddd4e8c6d] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 789.482869] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b5c531a2-a229-4998-9a9b-9ebb1e7afff2 tempest-ServerGroupTestJSON-1534054659 tempest-ServerGroupTestJSON-1534054659-project-member] Acquiring lock "5813dbfd-b575-455a-83e7-90b000a05ef8" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 789.483157] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b5c531a2-a229-4998-9a9b-9ebb1e7afff2 tempest-ServerGroupTestJSON-1534054659 tempest-ServerGroupTestJSON-1534054659-project-member] Lock "5813dbfd-b575-455a-83e7-90b000a05ef8" acquired by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 789.604593] env[63088]: DEBUG oslo_concurrency.lockutils [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 789.805739] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10c843b6-4dda-475e-baa3-dc94f97a54a3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.815723] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48534938-2ecf-4ef9-9b91-7de22f2f3698 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.866549] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e08cb8b9-635b-48f2-92a8-8678f72f0fc9 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.876761] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e03ac3b8-6612-4540-82a8-2e6a4809f201 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.891553] env[63088]: DEBUG nova.compute.provider_tree [None req-96c00a75-0383-4506-a4c6-d540aa18aaef tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 789.931051] env[63088]: INFO nova.scheduler.client.report [None req-1c0c7b6c-8782-4a52-b8ee-b7fd4463c345 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Deleted allocations for instance fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3 [ 789.989032] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b5c531a2-a229-4998-9a9b-9ebb1e7afff2 tempest-ServerGroupTestJSON-1534054659 tempest-ServerGroupTestJSON-1534054659-project-member] Lock "5813dbfd-b575-455a-83e7-90b000a05ef8" "released" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: held 0.506s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 789.989888] env[63088]: DEBUG nova.compute.manager [None req-b5c531a2-a229-4998-9a9b-9ebb1e7afff2 tempest-ServerGroupTestJSON-1534054659 tempest-ServerGroupTestJSON-1534054659-project-member] [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 790.401577] env[63088]: DEBUG nova.scheduler.client.report [None req-96c00a75-0383-4506-a4c6-d540aa18aaef tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 790.446328] env[63088]: DEBUG oslo_concurrency.lockutils [None req-1c0c7b6c-8782-4a52-b8ee-b7fd4463c345 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Lock "fcdb6c28-4ec1-4d69-8fc1-9c1aa05521b3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 133.101s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 790.499132] env[63088]: DEBUG nova.compute.utils [None req-b5c531a2-a229-4998-9a9b-9ebb1e7afff2 tempest-ServerGroupTestJSON-1534054659 tempest-ServerGroupTestJSON-1534054659-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 790.505043] env[63088]: DEBUG nova.compute.manager [None req-b5c531a2-a229-4998-9a9b-9ebb1e7afff2 tempest-ServerGroupTestJSON-1534054659 tempest-ServerGroupTestJSON-1534054659-project-member] [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 790.505043] env[63088]: DEBUG nova.network.neutron [None req-b5c531a2-a229-4998-9a9b-9ebb1e7afff2 tempest-ServerGroupTestJSON-1534054659 tempest-ServerGroupTestJSON-1534054659-project-member] [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 790.585319] env[63088]: DEBUG nova.policy [None req-b5c531a2-a229-4998-9a9b-9ebb1e7afff2 tempest-ServerGroupTestJSON-1534054659 tempest-ServerGroupTestJSON-1534054659-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fcbb927a3b2c4593998fbaac8a95dbd9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1aac7ee698304a888bba0e468c6d11ef', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 790.906167] env[63088]: DEBUG oslo_concurrency.lockutils [None req-96c00a75-0383-4506-a4c6-d540aa18aaef tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.931s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 790.907172] env[63088]: ERROR nova.compute.manager [None req-96c00a75-0383-4506-a4c6-d540aa18aaef tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5d70ba36-7221-4886-8ae1-21223c696acf, please check neutron logs for more information. [ 790.907172] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] Traceback (most recent call last): [ 790.907172] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 790.907172] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] self.driver.spawn(context, instance, image_meta, [ 790.907172] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 790.907172] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] self._vmops.spawn(context, instance, image_meta, injected_files, [ 790.907172] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 790.907172] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] vm_ref = self.build_virtual_machine(instance, [ 790.907172] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 790.907172] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] vif_infos = vmwarevif.get_vif_info(self._session, [ 790.907172] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 790.907863] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] for vif in network_info: [ 790.907863] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 790.907863] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] return self._sync_wrapper(fn, *args, **kwargs) [ 790.907863] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 790.907863] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] self.wait() [ 790.907863] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 790.907863] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] self[:] = self._gt.wait() [ 790.907863] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 790.907863] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] return self._exit_event.wait() [ 790.907863] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 790.907863] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] current.throw(*self._exc) [ 790.907863] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 790.907863] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] result = function(*args, **kwargs) [ 790.909015] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 790.909015] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] return func(*args, **kwargs) [ 790.909015] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 790.909015] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] raise e [ 790.909015] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 790.909015] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] nwinfo = self.network_api.allocate_for_instance( [ 790.909015] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 790.909015] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] created_port_ids = self._update_ports_for_instance( [ 790.909015] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 790.909015] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] with excutils.save_and_reraise_exception(): [ 790.909015] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 790.909015] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] self.force_reraise() [ 790.909015] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 790.909557] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] raise self.value [ 790.909557] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 790.909557] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] updated_port = self._update_port( [ 790.909557] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 790.909557] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] _ensure_no_port_binding_failure(port) [ 790.909557] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 790.909557] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] raise exception.PortBindingFailed(port_id=port['id']) [ 790.909557] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] nova.exception.PortBindingFailed: Binding failed for port 5d70ba36-7221-4886-8ae1-21223c696acf, please check neutron logs for more information. [ 790.909557] env[63088]: ERROR nova.compute.manager [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] [ 790.909557] env[63088]: DEBUG nova.compute.utils [None req-96c00a75-0383-4506-a4c6-d540aa18aaef tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] Binding failed for port 5d70ba36-7221-4886-8ae1-21223c696acf, please check neutron logs for more information. {{(pid=63088) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 790.911156] env[63088]: DEBUG nova.compute.manager [None req-96c00a75-0383-4506-a4c6-d540aa18aaef tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] Build of instance 87361d1c-e9cd-4177-bf3d-5d7551192073 was re-scheduled: Binding failed for port 5d70ba36-7221-4886-8ae1-21223c696acf, please check neutron logs for more information. {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 790.911492] env[63088]: DEBUG nova.compute.manager [None req-96c00a75-0383-4506-a4c6-d540aa18aaef tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] Unplugging VIFs for instance {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 790.911575] env[63088]: DEBUG oslo_concurrency.lockutils [None req-96c00a75-0383-4506-a4c6-d540aa18aaef tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Acquiring lock "refresh_cache-87361d1c-e9cd-4177-bf3d-5d7551192073" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 790.913331] env[63088]: DEBUG oslo_concurrency.lockutils [None req-96c00a75-0383-4506-a4c6-d540aa18aaef tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Acquired lock "refresh_cache-87361d1c-e9cd-4177-bf3d-5d7551192073" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 790.913331] env[63088]: DEBUG nova.network.neutron [None req-96c00a75-0383-4506-a4c6-d540aa18aaef tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 790.916495] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 13.415s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 790.951866] env[63088]: DEBUG nova.compute.manager [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 791.005990] env[63088]: DEBUG nova.compute.manager [None req-b5c531a2-a229-4998-9a9b-9ebb1e7afff2 tempest-ServerGroupTestJSON-1534054659 tempest-ServerGroupTestJSON-1534054659-project-member] [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 791.183512] env[63088]: DEBUG nova.network.neutron [None req-b5c531a2-a229-4998-9a9b-9ebb1e7afff2 tempest-ServerGroupTestJSON-1534054659 tempest-ServerGroupTestJSON-1534054659-project-member] [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] Successfully created port: 3235d186-bfbf-41f3-88af-330b3981042d {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 791.215063] env[63088]: DEBUG oslo_concurrency.lockutils [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Acquiring lock "00925f94-dbf9-453c-a124-d8434679aedf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 791.215063] env[63088]: DEBUG oslo_concurrency.lockutils [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Lock "00925f94-dbf9-453c-a124-d8434679aedf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 791.239635] env[63088]: DEBUG oslo_concurrency.lockutils [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Acquiring lock "7f341d80-450a-47b8-b26d-15f2ce3e378c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 791.239635] env[63088]: DEBUG oslo_concurrency.lockutils [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Lock "7f341d80-450a-47b8-b26d-15f2ce3e378c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 791.441206] env[63088]: DEBUG nova.network.neutron [None req-96c00a75-0383-4506-a4c6-d540aa18aaef tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 791.481860] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 791.523509] env[63088]: DEBUG nova.network.neutron [None req-96c00a75-0383-4506-a4c6-d540aa18aaef tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 792.020013] env[63088]: DEBUG nova.compute.manager [None req-b5c531a2-a229-4998-9a9b-9ebb1e7afff2 tempest-ServerGroupTestJSON-1534054659 tempest-ServerGroupTestJSON-1534054659-project-member] [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 792.030806] env[63088]: DEBUG oslo_concurrency.lockutils [None req-96c00a75-0383-4506-a4c6-d540aa18aaef tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Releasing lock "refresh_cache-87361d1c-e9cd-4177-bf3d-5d7551192073" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 792.033678] env[63088]: DEBUG nova.compute.manager [None req-96c00a75-0383-4506-a4c6-d540aa18aaef tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 792.033678] env[63088]: DEBUG nova.compute.manager [None req-96c00a75-0383-4506-a4c6-d540aa18aaef tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 792.033678] env[63088]: DEBUG nova.network.neutron [None req-96c00a75-0383-4506-a4c6-d540aa18aaef tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 792.054495] env[63088]: DEBUG nova.virt.hardware [None req-b5c531a2-a229-4998-9a9b-9ebb1e7afff2 tempest-ServerGroupTestJSON-1534054659 tempest-ServerGroupTestJSON-1534054659-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 792.055092] env[63088]: DEBUG nova.virt.hardware [None req-b5c531a2-a229-4998-9a9b-9ebb1e7afff2 tempest-ServerGroupTestJSON-1534054659 tempest-ServerGroupTestJSON-1534054659-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 792.055092] env[63088]: DEBUG nova.virt.hardware [None req-b5c531a2-a229-4998-9a9b-9ebb1e7afff2 tempest-ServerGroupTestJSON-1534054659 tempest-ServerGroupTestJSON-1534054659-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 792.055240] env[63088]: DEBUG nova.virt.hardware [None req-b5c531a2-a229-4998-9a9b-9ebb1e7afff2 tempest-ServerGroupTestJSON-1534054659 tempest-ServerGroupTestJSON-1534054659-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 792.055418] env[63088]: DEBUG nova.virt.hardware [None req-b5c531a2-a229-4998-9a9b-9ebb1e7afff2 tempest-ServerGroupTestJSON-1534054659 tempest-ServerGroupTestJSON-1534054659-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 792.055609] env[63088]: DEBUG nova.virt.hardware [None req-b5c531a2-a229-4998-9a9b-9ebb1e7afff2 tempest-ServerGroupTestJSON-1534054659 tempest-ServerGroupTestJSON-1534054659-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 792.056331] env[63088]: DEBUG nova.virt.hardware [None req-b5c531a2-a229-4998-9a9b-9ebb1e7afff2 tempest-ServerGroupTestJSON-1534054659 tempest-ServerGroupTestJSON-1534054659-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 792.056331] env[63088]: DEBUG nova.virt.hardware [None req-b5c531a2-a229-4998-9a9b-9ebb1e7afff2 tempest-ServerGroupTestJSON-1534054659 tempest-ServerGroupTestJSON-1534054659-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 792.056331] env[63088]: DEBUG nova.virt.hardware [None req-b5c531a2-a229-4998-9a9b-9ebb1e7afff2 tempest-ServerGroupTestJSON-1534054659 tempest-ServerGroupTestJSON-1534054659-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 792.056331] env[63088]: DEBUG nova.virt.hardware [None req-b5c531a2-a229-4998-9a9b-9ebb1e7afff2 tempest-ServerGroupTestJSON-1534054659 tempest-ServerGroupTestJSON-1534054659-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 792.056526] env[63088]: DEBUG nova.virt.hardware [None req-b5c531a2-a229-4998-9a9b-9ebb1e7afff2 tempest-ServerGroupTestJSON-1534054659 tempest-ServerGroupTestJSON-1534054659-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 792.059041] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ed5202c-8c4a-461d-8f40-a5d64c22387b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.061731] env[63088]: DEBUG nova.network.neutron [None req-96c00a75-0383-4506-a4c6-d540aa18aaef tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 792.072030] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78c18503-e3fa-4f4b-82c6-9163942aaf1d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.095238] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Acquiring lock "8b564e29-96cf-4abf-963d-142b413fb464" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 792.095577] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Lock "8b564e29-96cf-4abf-963d-142b413fb464" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 792.368165] env[63088]: DEBUG nova.compute.manager [req-50531559-b7ab-41d7-9b85-8d4dd33785fe req-b18b5f53-63de-47da-a922-31ee2f5a46da service nova] [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] Received event network-changed-3235d186-bfbf-41f3-88af-330b3981042d {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 792.368361] env[63088]: DEBUG nova.compute.manager [req-50531559-b7ab-41d7-9b85-8d4dd33785fe req-b18b5f53-63de-47da-a922-31ee2f5a46da service nova] [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] Refreshing instance network info cache due to event network-changed-3235d186-bfbf-41f3-88af-330b3981042d. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 792.368560] env[63088]: DEBUG oslo_concurrency.lockutils [req-50531559-b7ab-41d7-9b85-8d4dd33785fe req-b18b5f53-63de-47da-a922-31ee2f5a46da service nova] Acquiring lock "refresh_cache-53ea5c01-4be2-44a8-b74f-0d74031dfdd7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 792.371722] env[63088]: DEBUG oslo_concurrency.lockutils [req-50531559-b7ab-41d7-9b85-8d4dd33785fe req-b18b5f53-63de-47da-a922-31ee2f5a46da service nova] Acquired lock "refresh_cache-53ea5c01-4be2-44a8-b74f-0d74031dfdd7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 792.372368] env[63088]: DEBUG nova.network.neutron [req-50531559-b7ab-41d7-9b85-8d4dd33785fe req-b18b5f53-63de-47da-a922-31ee2f5a46da service nova] [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] Refreshing network info cache for port 3235d186-bfbf-41f3-88af-330b3981042d {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 792.460319] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance 87361d1c-e9cd-4177-bf3d-5d7551192073 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 792.460638] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance e6a72b08-38b6-44ae-8911-6b39d50ee1c3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 792.460638] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance 53ea5c01-4be2-44a8-b74f-0d74031dfdd7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 792.519319] env[63088]: ERROR nova.compute.manager [None req-b5c531a2-a229-4998-9a9b-9ebb1e7afff2 tempest-ServerGroupTestJSON-1534054659 tempest-ServerGroupTestJSON-1534054659-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3235d186-bfbf-41f3-88af-330b3981042d, please check neutron logs for more information. [ 792.519319] env[63088]: ERROR nova.compute.manager Traceback (most recent call last): [ 792.519319] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 792.519319] env[63088]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 792.519319] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 792.519319] env[63088]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 792.519319] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 792.519319] env[63088]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 792.519319] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 792.519319] env[63088]: ERROR nova.compute.manager self.force_reraise() [ 792.519319] env[63088]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 792.519319] env[63088]: ERROR nova.compute.manager raise self.value [ 792.519319] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 792.519319] env[63088]: ERROR nova.compute.manager updated_port = self._update_port( [ 792.519319] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 792.519319] env[63088]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 792.519969] env[63088]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 792.519969] env[63088]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 792.519969] env[63088]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3235d186-bfbf-41f3-88af-330b3981042d, please check neutron logs for more information. [ 792.519969] env[63088]: ERROR nova.compute.manager [ 792.519969] env[63088]: Traceback (most recent call last): [ 792.519969] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 792.519969] env[63088]: listener.cb(fileno) [ 792.519969] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 792.519969] env[63088]: result = function(*args, **kwargs) [ 792.519969] env[63088]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 792.519969] env[63088]: return func(*args, **kwargs) [ 792.519969] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 792.519969] env[63088]: raise e [ 792.519969] env[63088]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 792.519969] env[63088]: nwinfo = self.network_api.allocate_for_instance( [ 792.519969] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 792.519969] env[63088]: created_port_ids = self._update_ports_for_instance( [ 792.519969] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 792.519969] env[63088]: with excutils.save_and_reraise_exception(): [ 792.519969] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 792.519969] env[63088]: self.force_reraise() [ 792.519969] env[63088]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 792.519969] env[63088]: raise self.value [ 792.519969] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 792.519969] env[63088]: updated_port = self._update_port( [ 792.519969] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 792.519969] env[63088]: _ensure_no_port_binding_failure(port) [ 792.519969] env[63088]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 792.519969] env[63088]: raise exception.PortBindingFailed(port_id=port['id']) [ 792.520958] env[63088]: nova.exception.PortBindingFailed: Binding failed for port 3235d186-bfbf-41f3-88af-330b3981042d, please check neutron logs for more information. [ 792.520958] env[63088]: Removing descriptor: 17 [ 792.520958] env[63088]: ERROR nova.compute.manager [None req-b5c531a2-a229-4998-9a9b-9ebb1e7afff2 tempest-ServerGroupTestJSON-1534054659 tempest-ServerGroupTestJSON-1534054659-project-member] [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3235d186-bfbf-41f3-88af-330b3981042d, please check neutron logs for more information. [ 792.520958] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] Traceback (most recent call last): [ 792.520958] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 792.520958] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] yield resources [ 792.520958] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 792.520958] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] self.driver.spawn(context, instance, image_meta, [ 792.520958] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 792.520958] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 792.520958] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 792.520958] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] vm_ref = self.build_virtual_machine(instance, [ 792.521402] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 792.521402] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] vif_infos = vmwarevif.get_vif_info(self._session, [ 792.521402] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 792.521402] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] for vif in network_info: [ 792.521402] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 792.521402] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] return self._sync_wrapper(fn, *args, **kwargs) [ 792.521402] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 792.521402] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] self.wait() [ 792.521402] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 792.521402] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] self[:] = self._gt.wait() [ 792.521402] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 792.521402] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] return self._exit_event.wait() [ 792.521402] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 792.521838] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] result = hub.switch() [ 792.521838] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 792.521838] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] return self.greenlet.switch() [ 792.521838] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 792.521838] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] result = function(*args, **kwargs) [ 792.521838] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 792.521838] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] return func(*args, **kwargs) [ 792.521838] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 792.521838] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] raise e [ 792.521838] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 792.521838] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] nwinfo = self.network_api.allocate_for_instance( [ 792.521838] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 792.521838] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] created_port_ids = self._update_ports_for_instance( [ 792.522237] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 792.522237] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] with excutils.save_and_reraise_exception(): [ 792.522237] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 792.522237] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] self.force_reraise() [ 792.522237] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 792.522237] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] raise self.value [ 792.522237] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 792.522237] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] updated_port = self._update_port( [ 792.522237] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 792.522237] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] _ensure_no_port_binding_failure(port) [ 792.522237] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 792.522237] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] raise exception.PortBindingFailed(port_id=port['id']) [ 792.522604] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] nova.exception.PortBindingFailed: Binding failed for port 3235d186-bfbf-41f3-88af-330b3981042d, please check neutron logs for more information. [ 792.522604] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] [ 792.522604] env[63088]: INFO nova.compute.manager [None req-b5c531a2-a229-4998-9a9b-9ebb1e7afff2 tempest-ServerGroupTestJSON-1534054659 tempest-ServerGroupTestJSON-1534054659-project-member] [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] Terminating instance [ 792.522779] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b5c531a2-a229-4998-9a9b-9ebb1e7afff2 tempest-ServerGroupTestJSON-1534054659 tempest-ServerGroupTestJSON-1534054659-project-member] Acquiring lock "refresh_cache-53ea5c01-4be2-44a8-b74f-0d74031dfdd7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 792.565341] env[63088]: DEBUG nova.network.neutron [None req-96c00a75-0383-4506-a4c6-d540aa18aaef tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 792.890606] env[63088]: DEBUG nova.network.neutron [req-50531559-b7ab-41d7-9b85-8d4dd33785fe req-b18b5f53-63de-47da-a922-31ee2f5a46da service nova] [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 792.964357] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance db032da3-d334-486f-a428-af1c8a3c360a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 792.989628] env[63088]: DEBUG nova.network.neutron [req-50531559-b7ab-41d7-9b85-8d4dd33785fe req-b18b5f53-63de-47da-a922-31ee2f5a46da service nova] [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 793.068013] env[63088]: INFO nova.compute.manager [None req-96c00a75-0383-4506-a4c6-d540aa18aaef tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: 87361d1c-e9cd-4177-bf3d-5d7551192073] Took 1.03 seconds to deallocate network for instance. [ 793.466932] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance 11ec987f-9908-4366-81ae-cbfe4de67b2d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 793.492794] env[63088]: DEBUG oslo_concurrency.lockutils [req-50531559-b7ab-41d7-9b85-8d4dd33785fe req-b18b5f53-63de-47da-a922-31ee2f5a46da service nova] Releasing lock "refresh_cache-53ea5c01-4be2-44a8-b74f-0d74031dfdd7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 793.493206] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b5c531a2-a229-4998-9a9b-9ebb1e7afff2 tempest-ServerGroupTestJSON-1534054659 tempest-ServerGroupTestJSON-1534054659-project-member] Acquired lock "refresh_cache-53ea5c01-4be2-44a8-b74f-0d74031dfdd7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 793.493440] env[63088]: DEBUG nova.network.neutron [None req-b5c531a2-a229-4998-9a9b-9ebb1e7afff2 tempest-ServerGroupTestJSON-1534054659 tempest-ServerGroupTestJSON-1534054659-project-member] [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 793.970237] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance 4971b24c-6710-4f50-9846-727dad264b1f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 794.013451] env[63088]: DEBUG nova.network.neutron [None req-b5c531a2-a229-4998-9a9b-9ebb1e7afff2 tempest-ServerGroupTestJSON-1534054659 tempest-ServerGroupTestJSON-1534054659-project-member] [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 794.103480] env[63088]: INFO nova.scheduler.client.report [None req-96c00a75-0383-4506-a4c6-d540aa18aaef tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Deleted allocations for instance 87361d1c-e9cd-4177-bf3d-5d7551192073 [ 794.164321] env[63088]: DEBUG nova.network.neutron [None req-b5c531a2-a229-4998-9a9b-9ebb1e7afff2 tempest-ServerGroupTestJSON-1534054659 tempest-ServerGroupTestJSON-1534054659-project-member] [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 794.395235] env[63088]: DEBUG nova.compute.manager [req-31ca15e5-ebe6-4511-8c7d-02d549825334 req-79c51e5f-16c4-43b3-986c-e85c3cfc809f service nova] [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] Received event network-vif-deleted-3235d186-bfbf-41f3-88af-330b3981042d {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 794.473841] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance 4847f15e-98d4-401c-91ff-067e84a85727 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 794.612906] env[63088]: DEBUG oslo_concurrency.lockutils [None req-96c00a75-0383-4506-a4c6-d540aa18aaef tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Lock "87361d1c-e9cd-4177-bf3d-5d7551192073" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 134.348s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 794.666298] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b5c531a2-a229-4998-9a9b-9ebb1e7afff2 tempest-ServerGroupTestJSON-1534054659 tempest-ServerGroupTestJSON-1534054659-project-member] Releasing lock "refresh_cache-53ea5c01-4be2-44a8-b74f-0d74031dfdd7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 794.666767] env[63088]: DEBUG nova.compute.manager [None req-b5c531a2-a229-4998-9a9b-9ebb1e7afff2 tempest-ServerGroupTestJSON-1534054659 tempest-ServerGroupTestJSON-1534054659-project-member] [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 794.666972] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-b5c531a2-a229-4998-9a9b-9ebb1e7afff2 tempest-ServerGroupTestJSON-1534054659 tempest-ServerGroupTestJSON-1534054659-project-member] [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 794.667572] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6d64d73c-10d3-400b-b1b3-d04f89a052fa {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.676605] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1c80fe9-372b-4bd2-950b-62741ffaf11f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.700312] env[63088]: WARNING nova.virt.vmwareapi.vmops [None req-b5c531a2-a229-4998-9a9b-9ebb1e7afff2 tempest-ServerGroupTestJSON-1534054659 tempest-ServerGroupTestJSON-1534054659-project-member] [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 53ea5c01-4be2-44a8-b74f-0d74031dfdd7 could not be found. [ 794.700557] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-b5c531a2-a229-4998-9a9b-9ebb1e7afff2 tempest-ServerGroupTestJSON-1534054659 tempest-ServerGroupTestJSON-1534054659-project-member] [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 794.700738] env[63088]: INFO nova.compute.manager [None req-b5c531a2-a229-4998-9a9b-9ebb1e7afff2 tempest-ServerGroupTestJSON-1534054659 tempest-ServerGroupTestJSON-1534054659-project-member] [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] Took 0.03 seconds to destroy the instance on the hypervisor. [ 794.704039] env[63088]: DEBUG oslo.service.loopingcall [None req-b5c531a2-a229-4998-9a9b-9ebb1e7afff2 tempest-ServerGroupTestJSON-1534054659 tempest-ServerGroupTestJSON-1534054659-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 794.704039] env[63088]: DEBUG nova.compute.manager [-] [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 794.704039] env[63088]: DEBUG nova.network.neutron [-] [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 794.724773] env[63088]: DEBUG nova.network.neutron [-] [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 794.980023] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance e845f60b-004f-4a5c-813e-0c183e99eba6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 795.118014] env[63088]: DEBUG nova.compute.manager [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] [instance: b159649b-ebd5-4c7d-9074-dc0e6395a947] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 795.229871] env[63088]: DEBUG nova.network.neutron [-] [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 795.486773] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance 4f6e2d99-7300-42e6-8abf-362ddd4e8c6d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 795.649890] env[63088]: DEBUG oslo_concurrency.lockutils [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 795.738033] env[63088]: INFO nova.compute.manager [-] [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] Took 1.03 seconds to deallocate network for instance. [ 795.740040] env[63088]: DEBUG nova.compute.claims [None req-b5c531a2-a229-4998-9a9b-9ebb1e7afff2 tempest-ServerGroupTestJSON-1534054659 tempest-ServerGroupTestJSON-1534054659-project-member] [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] Aborting claim: {{(pid=63088) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 795.740307] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b5c531a2-a229-4998-9a9b-9ebb1e7afff2 tempest-ServerGroupTestJSON-1534054659 tempest-ServerGroupTestJSON-1534054659-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 795.990715] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance 07fd3e6f-e0af-467c-9039-238be5c58d25 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 796.495431] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance b159649b-ebd5-4c7d-9074-dc0e6395a947 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 796.999056] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance 34ce411f-40c8-446e-b685-cd3ce07663dd has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 797.231902] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquiring lock "7b6aadb7-e34b-42b7-b69f-370434f5b665" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 797.232166] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lock "7b6aadb7-e34b-42b7-b69f-370434f5b665" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 797.503461] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance 1d421bf0-925d-408c-a612-b659da92799f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 797.600126] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Acquiring lock "dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 797.600368] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Lock "dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 798.006844] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance 2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 798.509274] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance 00925f94-dbf9-453c-a124-d8434679aedf has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 799.013082] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance 7f341d80-450a-47b8-b26d-15f2ce3e378c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 799.516674] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance 8b564e29-96cf-4abf-963d-142b413fb464 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 799.516994] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Total usable vcpus: 48, total allocated vcpus: 2 {{(pid=63088) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 799.517083] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=896MB phys_disk=200GB used_disk=2GB total_vcpus=48 used_vcpus=2 pci_stats=[] {{(pid=63088) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 799.731214] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d605597-8996-4646-aab5-c859bf112832 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.738791] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f156a514-e410-43c2-a8fa-df85ade9a162 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.768741] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83ea33a1-8f80-46b5-b548-797862de18d8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.775649] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f638c7d5-9913-40ea-9980-8a2fb25b41d4 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.788700] env[63088]: DEBUG nova.compute.provider_tree [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 800.291359] env[63088]: DEBUG nova.scheduler.client.report [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 800.796471] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63088) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 800.796810] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 9.880s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 800.796984] env[63088]: DEBUG oslo_concurrency.lockutils [None req-02aa52fc-420f-46cd-9829-a68800c429e2 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 22.763s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 800.799893] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 800.800870] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Cleaning up deleted instances {{(pid=63088) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11226}} [ 801.306033] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] There are 2 instances to clean {{(pid=63088) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11235}} [ 801.306314] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: c4777721-3f65-455f-9973-c1ed0732de34] Instance has had 0 of 5 cleanup attempts {{(pid=63088) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 801.547735] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf2ea14c-7953-43de-87d5-5b622343c985 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.555415] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e6bcbed-e713-4194-9ca9-d9b376656dc1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.585241] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ad9a478-446c-443f-8362-16e6ca5e0e80 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.592862] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c90e23ad-c751-40d7-9078-c93849d705eb {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.606535] env[63088]: DEBUG nova.compute.provider_tree [None req-02aa52fc-420f-46cd-9829-a68800c429e2 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 801.809758] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: e8cfbce4-cfb2-463e-9f38-78a058dab6e1] Instance has had 0 of 5 cleanup attempts {{(pid=63088) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 802.109129] env[63088]: DEBUG nova.scheduler.client.report [None req-02aa52fc-420f-46cd-9829-a68800c429e2 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 802.312783] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 802.312958] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Cleaning up deleted instances with incomplete migration {{(pid=63088) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11264}} [ 802.615732] env[63088]: DEBUG oslo_concurrency.lockutils [None req-02aa52fc-420f-46cd-9829-a68800c429e2 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.819s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 802.616386] env[63088]: ERROR nova.compute.manager [None req-02aa52fc-420f-46cd-9829-a68800c429e2 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 7876ef7c-e226-4a28-ac65-0d3647d7d12c, please check neutron logs for more information. [ 802.616386] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] Traceback (most recent call last): [ 802.616386] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 802.616386] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] self.driver.spawn(context, instance, image_meta, [ 802.616386] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 802.616386] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 802.616386] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 802.616386] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] vm_ref = self.build_virtual_machine(instance, [ 802.616386] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 802.616386] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] vif_infos = vmwarevif.get_vif_info(self._session, [ 802.616386] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 802.616722] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] for vif in network_info: [ 802.616722] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 802.616722] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] return self._sync_wrapper(fn, *args, **kwargs) [ 802.616722] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 802.616722] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] self.wait() [ 802.616722] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 802.616722] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] self[:] = self._gt.wait() [ 802.616722] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 802.616722] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] return self._exit_event.wait() [ 802.616722] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 802.616722] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] current.throw(*self._exc) [ 802.616722] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 802.616722] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] result = function(*args, **kwargs) [ 802.616988] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 802.616988] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] return func(*args, **kwargs) [ 802.616988] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 802.616988] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] raise e [ 802.616988] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 802.616988] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] nwinfo = self.network_api.allocate_for_instance( [ 802.616988] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 802.616988] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] created_port_ids = self._update_ports_for_instance( [ 802.616988] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 802.616988] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] with excutils.save_and_reraise_exception(): [ 802.616988] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 802.616988] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] self.force_reraise() [ 802.616988] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 802.617271] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] raise self.value [ 802.617271] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 802.617271] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] updated_port = self._update_port( [ 802.617271] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 802.617271] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] _ensure_no_port_binding_failure(port) [ 802.617271] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 802.617271] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] raise exception.PortBindingFailed(port_id=port['id']) [ 802.617271] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] nova.exception.PortBindingFailed: Binding failed for port 7876ef7c-e226-4a28-ac65-0d3647d7d12c, please check neutron logs for more information. [ 802.617271] env[63088]: ERROR nova.compute.manager [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] [ 802.617271] env[63088]: DEBUG nova.compute.utils [None req-02aa52fc-420f-46cd-9829-a68800c429e2 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] Binding failed for port 7876ef7c-e226-4a28-ac65-0d3647d7d12c, please check neutron logs for more information. {{(pid=63088) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 802.618563] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.281s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 802.619867] env[63088]: INFO nova.compute.claims [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: db032da3-d334-486f-a428-af1c8a3c360a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 802.622416] env[63088]: DEBUG nova.compute.manager [None req-02aa52fc-420f-46cd-9829-a68800c429e2 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] Build of instance e6a72b08-38b6-44ae-8911-6b39d50ee1c3 was re-scheduled: Binding failed for port 7876ef7c-e226-4a28-ac65-0d3647d7d12c, please check neutron logs for more information. {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 802.622823] env[63088]: DEBUG nova.compute.manager [None req-02aa52fc-420f-46cd-9829-a68800c429e2 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] Unplugging VIFs for instance {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 802.623041] env[63088]: DEBUG oslo_concurrency.lockutils [None req-02aa52fc-420f-46cd-9829-a68800c429e2 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Acquiring lock "refresh_cache-e6a72b08-38b6-44ae-8911-6b39d50ee1c3" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 802.623193] env[63088]: DEBUG oslo_concurrency.lockutils [None req-02aa52fc-420f-46cd-9829-a68800c429e2 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Acquired lock "refresh_cache-e6a72b08-38b6-44ae-8911-6b39d50ee1c3" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 802.623352] env[63088]: DEBUG nova.network.neutron [None req-02aa52fc-420f-46cd-9829-a68800c429e2 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 802.816108] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 803.142178] env[63088]: DEBUG nova.network.neutron [None req-02aa52fc-420f-46cd-9829-a68800c429e2 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 803.198489] env[63088]: DEBUG nova.network.neutron [None req-02aa52fc-420f-46cd-9829-a68800c429e2 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 803.700734] env[63088]: DEBUG oslo_concurrency.lockutils [None req-02aa52fc-420f-46cd-9829-a68800c429e2 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Releasing lock "refresh_cache-e6a72b08-38b6-44ae-8911-6b39d50ee1c3" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 803.701458] env[63088]: DEBUG nova.compute.manager [None req-02aa52fc-420f-46cd-9829-a68800c429e2 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 803.701458] env[63088]: DEBUG nova.compute.manager [None req-02aa52fc-420f-46cd-9829-a68800c429e2 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 803.701458] env[63088]: DEBUG nova.network.neutron [None req-02aa52fc-420f-46cd-9829-a68800c429e2 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 803.719086] env[63088]: DEBUG nova.network.neutron [None req-02aa52fc-420f-46cd-9829-a68800c429e2 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 803.989966] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99fef632-7c7b-4cc0-88e8-06536270c40a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.997482] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8939267-f60f-45a9-a901-b5607abfe703 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.026237] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99c823a1-f83c-4957-9c29-e7303ec015c1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.034379] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d04cc048-81de-4032-8db5-0dfd08ea4eed {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.047144] env[63088]: DEBUG nova.compute.provider_tree [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 804.222014] env[63088]: DEBUG nova.network.neutron [None req-02aa52fc-420f-46cd-9829-a68800c429e2 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 804.550066] env[63088]: DEBUG nova.scheduler.client.report [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 804.724991] env[63088]: INFO nova.compute.manager [None req-02aa52fc-420f-46cd-9829-a68800c429e2 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6a72b08-38b6-44ae-8911-6b39d50ee1c3] Took 1.02 seconds to deallocate network for instance. [ 805.055492] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.437s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 805.055813] env[63088]: DEBUG nova.compute.manager [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: db032da3-d334-486f-a428-af1c8a3c360a] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 805.058413] env[63088]: DEBUG oslo_concurrency.lockutils [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.914s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 805.059895] env[63088]: INFO nova.compute.claims [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 11ec987f-9908-4366-81ae-cbfe4de67b2d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 805.564539] env[63088]: DEBUG nova.compute.utils [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 805.567751] env[63088]: DEBUG nova.compute.manager [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: db032da3-d334-486f-a428-af1c8a3c360a] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 805.567918] env[63088]: DEBUG nova.network.neutron [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: db032da3-d334-486f-a428-af1c8a3c360a] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 805.621037] env[63088]: DEBUG nova.policy [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '98509947cc6745ef8037ff68da8925d2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '900466c3c09a4f928bdd4b602715c838', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 805.758275] env[63088]: INFO nova.scheduler.client.report [None req-02aa52fc-420f-46cd-9829-a68800c429e2 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Deleted allocations for instance e6a72b08-38b6-44ae-8911-6b39d50ee1c3 [ 805.924771] env[63088]: DEBUG nova.network.neutron [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: db032da3-d334-486f-a428-af1c8a3c360a] Successfully created port: 98219984-3198-4e54-8c93-024446f958a3 {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 806.068275] env[63088]: DEBUG nova.compute.manager [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: db032da3-d334-486f-a428-af1c8a3c360a] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 806.270161] env[63088]: DEBUG oslo_concurrency.lockutils [None req-02aa52fc-420f-46cd-9829-a68800c429e2 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Lock "e6a72b08-38b6-44ae-8911-6b39d50ee1c3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 122.892s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 806.312282] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a51b0934-37e2-4c02-bbf1-3c0d907df00f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.320338] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e10c0ed6-900b-4757-bafa-6ef4f95381c6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.351409] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99630df8-9344-402f-9cb1-373fbeb61181 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.359173] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bab900b-7bde-41e4-beea-10709630c2d5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.372150] env[63088]: DEBUG nova.compute.provider_tree [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 806.772593] env[63088]: DEBUG nova.compute.manager [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] [instance: 34ce411f-40c8-446e-b685-cd3ce07663dd] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 806.875294] env[63088]: DEBUG nova.scheduler.client.report [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 807.080736] env[63088]: DEBUG nova.compute.manager [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: db032da3-d334-486f-a428-af1c8a3c360a] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 807.106321] env[63088]: DEBUG nova.virt.hardware [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 807.106575] env[63088]: DEBUG nova.virt.hardware [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 807.107518] env[63088]: DEBUG nova.virt.hardware [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 807.107518] env[63088]: DEBUG nova.virt.hardware [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 807.107518] env[63088]: DEBUG nova.virt.hardware [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 807.107518] env[63088]: DEBUG nova.virt.hardware [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 807.107518] env[63088]: DEBUG nova.virt.hardware [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 807.107686] env[63088]: DEBUG nova.virt.hardware [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 807.107718] env[63088]: DEBUG nova.virt.hardware [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 807.107870] env[63088]: DEBUG nova.virt.hardware [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 807.108046] env[63088]: DEBUG nova.virt.hardware [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 807.108904] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d64a640-ccdc-4dab-b350-be0b82f15bee {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.116723] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1569d504-93de-46bb-9da1-c3de21a285dc {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.292322] env[63088]: DEBUG oslo_concurrency.lockutils [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 807.380622] env[63088]: DEBUG oslo_concurrency.lockutils [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.322s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 807.381166] env[63088]: DEBUG nova.compute.manager [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 11ec987f-9908-4366-81ae-cbfe4de67b2d] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 807.383707] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.387s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 807.385183] env[63088]: INFO nova.compute.claims [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 807.431553] env[63088]: DEBUG nova.compute.manager [req-14bb7f22-89aa-4c94-b912-0337ac7ba602 req-9ace4e47-c2ef-41d3-84c1-290814dcdcb1 service nova] [instance: db032da3-d334-486f-a428-af1c8a3c360a] Received event network-vif-plugged-98219984-3198-4e54-8c93-024446f958a3 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 807.431802] env[63088]: DEBUG oslo_concurrency.lockutils [req-14bb7f22-89aa-4c94-b912-0337ac7ba602 req-9ace4e47-c2ef-41d3-84c1-290814dcdcb1 service nova] Acquiring lock "db032da3-d334-486f-a428-af1c8a3c360a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 807.432009] env[63088]: DEBUG oslo_concurrency.lockutils [req-14bb7f22-89aa-4c94-b912-0337ac7ba602 req-9ace4e47-c2ef-41d3-84c1-290814dcdcb1 service nova] Lock "db032da3-d334-486f-a428-af1c8a3c360a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 807.432181] env[63088]: DEBUG oslo_concurrency.lockutils [req-14bb7f22-89aa-4c94-b912-0337ac7ba602 req-9ace4e47-c2ef-41d3-84c1-290814dcdcb1 service nova] Lock "db032da3-d334-486f-a428-af1c8a3c360a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 807.432344] env[63088]: DEBUG nova.compute.manager [req-14bb7f22-89aa-4c94-b912-0337ac7ba602 req-9ace4e47-c2ef-41d3-84c1-290814dcdcb1 service nova] [instance: db032da3-d334-486f-a428-af1c8a3c360a] No waiting events found dispatching network-vif-plugged-98219984-3198-4e54-8c93-024446f958a3 {{(pid=63088) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 807.432535] env[63088]: WARNING nova.compute.manager [req-14bb7f22-89aa-4c94-b912-0337ac7ba602 req-9ace4e47-c2ef-41d3-84c1-290814dcdcb1 service nova] [instance: db032da3-d334-486f-a428-af1c8a3c360a] Received unexpected event network-vif-plugged-98219984-3198-4e54-8c93-024446f958a3 for instance with vm_state building and task_state spawning. [ 807.522906] env[63088]: DEBUG nova.network.neutron [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: db032da3-d334-486f-a428-af1c8a3c360a] Successfully updated port: 98219984-3198-4e54-8c93-024446f958a3 {{(pid=63088) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 807.559648] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Acquiring lock "e6b0ce08-d67f-458f-92f5-1e904d03a5ef" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 807.560171] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Lock "e6b0ce08-d67f-458f-92f5-1e904d03a5ef" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 807.891624] env[63088]: DEBUG nova.compute.utils [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 807.892925] env[63088]: DEBUG nova.compute.manager [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 11ec987f-9908-4366-81ae-cbfe4de67b2d] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 807.893104] env[63088]: DEBUG nova.network.neutron [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 11ec987f-9908-4366-81ae-cbfe4de67b2d] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 807.931693] env[63088]: DEBUG nova.policy [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '139b65d9cf314b3db1d685d63bc19bba', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e48f62dee9ad4e9b94b67bd871db5c63', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 808.027969] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Acquiring lock "refresh_cache-db032da3-d334-486f-a428-af1c8a3c360a" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 808.028159] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Acquired lock "refresh_cache-db032da3-d334-486f-a428-af1c8a3c360a" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 808.028318] env[63088]: DEBUG nova.network.neutron [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: db032da3-d334-486f-a428-af1c8a3c360a] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 808.233209] env[63088]: DEBUG nova.network.neutron [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 11ec987f-9908-4366-81ae-cbfe4de67b2d] Successfully created port: 90ea719e-0a3d-4bae-a472-a58cbb3c1541 {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 808.396952] env[63088]: DEBUG nova.compute.manager [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 11ec987f-9908-4366-81ae-cbfe4de67b2d] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 808.561252] env[63088]: DEBUG nova.network.neutron [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: db032da3-d334-486f-a428-af1c8a3c360a] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 808.613895] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0948fe0d-495e-49e0-9ef7-775989d4952e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.623349] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2275192d-d120-452c-9630-6bda46b11968 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.653931] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b52358d-979b-47c3-b5d5-e3a0f3d28714 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.660659] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cd06482-974d-4060-ac11-221567cbc83e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.673186] env[63088]: DEBUG nova.compute.provider_tree [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 808.715033] env[63088]: DEBUG nova.network.neutron [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: db032da3-d334-486f-a428-af1c8a3c360a] Updating instance_info_cache with network_info: [{"id": "98219984-3198-4e54-8c93-024446f958a3", "address": "fa:16:3e:6d:e6:fe", "network": {"id": "7db9302e-1c7a-42dc-8084-98e2a57ad7bf", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-2073161169-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "900466c3c09a4f928bdd4b602715c838", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "87bbf4e0-9064-4516-b7e7-44973f817205", "external-id": "nsx-vlan-transportzone-507", "segmentation_id": 507, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap98219984-31", "ovs_interfaceid": "98219984-3198-4e54-8c93-024446f958a3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 809.175981] env[63088]: DEBUG nova.scheduler.client.report [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 809.217054] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Releasing lock "refresh_cache-db032da3-d334-486f-a428-af1c8a3c360a" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 809.217870] env[63088]: DEBUG nova.compute.manager [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: db032da3-d334-486f-a428-af1c8a3c360a] Instance network_info: |[{"id": "98219984-3198-4e54-8c93-024446f958a3", "address": "fa:16:3e:6d:e6:fe", "network": {"id": "7db9302e-1c7a-42dc-8084-98e2a57ad7bf", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-2073161169-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "900466c3c09a4f928bdd4b602715c838", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "87bbf4e0-9064-4516-b7e7-44973f817205", "external-id": "nsx-vlan-transportzone-507", "segmentation_id": 507, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap98219984-31", "ovs_interfaceid": "98219984-3198-4e54-8c93-024446f958a3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 809.218032] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: db032da3-d334-486f-a428-af1c8a3c360a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6d:e6:fe', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '87bbf4e0-9064-4516-b7e7-44973f817205', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '98219984-3198-4e54-8c93-024446f958a3', 'vif_model': 'vmxnet3'}] {{(pid=63088) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 809.225702] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Creating folder: Project (900466c3c09a4f928bdd4b602715c838). Parent ref: group-v275816. {{(pid=63088) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 809.226220] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6f98a10d-6361-4404-810f-231ae350c59b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.238529] env[63088]: INFO nova.virt.vmwareapi.vm_util [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Created folder: Project (900466c3c09a4f928bdd4b602715c838) in parent group-v275816. [ 809.238529] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Creating folder: Instances. Parent ref: group-v275833. {{(pid=63088) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 809.238529] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c801d0ee-5556-433b-9664-e4b259e6ba8f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.247674] env[63088]: INFO nova.virt.vmwareapi.vm_util [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Created folder: Instances in parent group-v275833. [ 809.247891] env[63088]: DEBUG oslo.service.loopingcall [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 809.248070] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: db032da3-d334-486f-a428-af1c8a3c360a] Creating VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 809.248264] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3afe1740-de6f-4b16-a5f1-db8540cc9b77 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.265906] env[63088]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 809.265906] env[63088]: value = "task-1284834" [ 809.265906] env[63088]: _type = "Task" [ 809.265906] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.273340] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1284834, 'name': CreateVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.406400] env[63088]: DEBUG nova.compute.manager [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 11ec987f-9908-4366-81ae-cbfe4de67b2d] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 809.434204] env[63088]: DEBUG nova.virt.hardware [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 809.434204] env[63088]: DEBUG nova.virt.hardware [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 809.434204] env[63088]: DEBUG nova.virt.hardware [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 809.434441] env[63088]: DEBUG nova.virt.hardware [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 809.434636] env[63088]: DEBUG nova.virt.hardware [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 809.434817] env[63088]: DEBUG nova.virt.hardware [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 809.435112] env[63088]: DEBUG nova.virt.hardware [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 809.435335] env[63088]: DEBUG nova.virt.hardware [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 809.435570] env[63088]: DEBUG nova.virt.hardware [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 809.435767] env[63088]: DEBUG nova.virt.hardware [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 809.437042] env[63088]: DEBUG nova.virt.hardware [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 809.437852] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-808799b5-90f3-4807-93d2-be0ac461efd6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.446657] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0fccae9-503f-40f9-ad8b-e27648f00958 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.466450] env[63088]: DEBUG nova.compute.manager [req-64f952b7-220d-4b51-93d0-0f7a8e84b912 req-7804b51b-e62c-41a7-93cb-ee94057461af service nova] [instance: db032da3-d334-486f-a428-af1c8a3c360a] Received event network-changed-98219984-3198-4e54-8c93-024446f958a3 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 809.466640] env[63088]: DEBUG nova.compute.manager [req-64f952b7-220d-4b51-93d0-0f7a8e84b912 req-7804b51b-e62c-41a7-93cb-ee94057461af service nova] [instance: db032da3-d334-486f-a428-af1c8a3c360a] Refreshing instance network info cache due to event network-changed-98219984-3198-4e54-8c93-024446f958a3. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 809.466850] env[63088]: DEBUG oslo_concurrency.lockutils [req-64f952b7-220d-4b51-93d0-0f7a8e84b912 req-7804b51b-e62c-41a7-93cb-ee94057461af service nova] Acquiring lock "refresh_cache-db032da3-d334-486f-a428-af1c8a3c360a" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 809.467029] env[63088]: DEBUG oslo_concurrency.lockutils [req-64f952b7-220d-4b51-93d0-0f7a8e84b912 req-7804b51b-e62c-41a7-93cb-ee94057461af service nova] Acquired lock "refresh_cache-db032da3-d334-486f-a428-af1c8a3c360a" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.467212] env[63088]: DEBUG nova.network.neutron [req-64f952b7-220d-4b51-93d0-0f7a8e84b912 req-7804b51b-e62c-41a7-93cb-ee94057461af service nova] [instance: db032da3-d334-486f-a428-af1c8a3c360a] Refreshing network info cache for port 98219984-3198-4e54-8c93-024446f958a3 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 809.680978] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.297s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 809.681502] env[63088]: DEBUG nova.compute.manager [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 809.686018] env[63088]: DEBUG oslo_concurrency.lockutils [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.820s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 809.686018] env[63088]: INFO nova.compute.claims [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 809.732423] env[63088]: DEBUG nova.network.neutron [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 11ec987f-9908-4366-81ae-cbfe4de67b2d] Successfully updated port: 90ea719e-0a3d-4bae-a472-a58cbb3c1541 {{(pid=63088) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 809.775072] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1284834, 'name': CreateVM_Task, 'duration_secs': 0.299418} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.775237] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: db032da3-d334-486f-a428-af1c8a3c360a] Created VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 809.782080] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 809.782252] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.782575] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 809.782811] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4dde2495-f4f7-4892-bc2c-dd70e15658be {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.787296] env[63088]: DEBUG oslo_vmware.api [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Waiting for the task: (returnval){ [ 809.787296] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]522add17-1009-ef36-c6d2-2d229cc4c5e4" [ 809.787296] env[63088]: _type = "Task" [ 809.787296] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.794307] env[63088]: DEBUG oslo_vmware.api [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]522add17-1009-ef36-c6d2-2d229cc4c5e4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.194025] env[63088]: DEBUG nova.compute.utils [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 810.195646] env[63088]: DEBUG nova.compute.manager [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 810.195827] env[63088]: DEBUG nova.network.neutron [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 810.234260] env[63088]: DEBUG oslo_concurrency.lockutils [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquiring lock "refresh_cache-11ec987f-9908-4366-81ae-cbfe4de67b2d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 810.234412] env[63088]: DEBUG oslo_concurrency.lockutils [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquired lock "refresh_cache-11ec987f-9908-4366-81ae-cbfe4de67b2d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 810.235809] env[63088]: DEBUG nova.network.neutron [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 11ec987f-9908-4366-81ae-cbfe4de67b2d] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 810.253031] env[63088]: DEBUG nova.network.neutron [req-64f952b7-220d-4b51-93d0-0f7a8e84b912 req-7804b51b-e62c-41a7-93cb-ee94057461af service nova] [instance: db032da3-d334-486f-a428-af1c8a3c360a] Updated VIF entry in instance network info cache for port 98219984-3198-4e54-8c93-024446f958a3. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 810.253397] env[63088]: DEBUG nova.network.neutron [req-64f952b7-220d-4b51-93d0-0f7a8e84b912 req-7804b51b-e62c-41a7-93cb-ee94057461af service nova] [instance: db032da3-d334-486f-a428-af1c8a3c360a] Updating instance_info_cache with network_info: [{"id": "98219984-3198-4e54-8c93-024446f958a3", "address": "fa:16:3e:6d:e6:fe", "network": {"id": "7db9302e-1c7a-42dc-8084-98e2a57ad7bf", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-2073161169-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "900466c3c09a4f928bdd4b602715c838", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "87bbf4e0-9064-4516-b7e7-44973f817205", "external-id": "nsx-vlan-transportzone-507", "segmentation_id": 507, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap98219984-31", "ovs_interfaceid": "98219984-3198-4e54-8c93-024446f958a3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 810.266212] env[63088]: DEBUG nova.policy [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '93f1d334e71a4a99ba300c3ee6d70243', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'aa3d24a1a6c0430985fd80365d986ee1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 810.302909] env[63088]: DEBUG oslo_vmware.api [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]522add17-1009-ef36-c6d2-2d229cc4c5e4, 'name': SearchDatastore_Task, 'duration_secs': 0.01334} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.304919] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 810.305180] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: db032da3-d334-486f-a428-af1c8a3c360a] Processing image 1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 810.305422] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 810.305545] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 810.305720] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 810.306157] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f2328f71-5cae-4d0b-87c7-10f578894b02 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.315434] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 810.315434] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 810.316155] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7ff82fe0-571e-4112-90d1-96b75d2f7e6a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.321360] env[63088]: DEBUG oslo_vmware.api [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Waiting for the task: (returnval){ [ 810.321360] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]5225d96f-3616-e022-e147-e5aa501fcd9e" [ 810.321360] env[63088]: _type = "Task" [ 810.321360] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.330504] env[63088]: DEBUG oslo_vmware.api [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]5225d96f-3616-e022-e147-e5aa501fcd9e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.599411] env[63088]: DEBUG nova.network.neutron [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Successfully created port: d5a10951-ae2e-48f5-b0ca-b1f144e5fe3a {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 810.696582] env[63088]: DEBUG nova.compute.manager [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 810.755931] env[63088]: DEBUG oslo_concurrency.lockutils [req-64f952b7-220d-4b51-93d0-0f7a8e84b912 req-7804b51b-e62c-41a7-93cb-ee94057461af service nova] Releasing lock "refresh_cache-db032da3-d334-486f-a428-af1c8a3c360a" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 810.771026] env[63088]: DEBUG nova.network.neutron [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 11ec987f-9908-4366-81ae-cbfe4de67b2d] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 810.832979] env[63088]: DEBUG oslo_vmware.api [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]5225d96f-3616-e022-e147-e5aa501fcd9e, 'name': SearchDatastore_Task, 'duration_secs': 0.008304} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.836119] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bca6e117-b79c-4ff9-b790-878683291789 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.843858] env[63088]: DEBUG oslo_vmware.api [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Waiting for the task: (returnval){ [ 810.843858] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52f23935-a419-5477-bb2a-8d5017c2dc3a" [ 810.843858] env[63088]: _type = "Task" [ 810.843858] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.856261] env[63088]: DEBUG oslo_vmware.api [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52f23935-a419-5477-bb2a-8d5017c2dc3a, 'name': SearchDatastore_Task, 'duration_secs': 0.008936} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.856261] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 810.856261] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] db032da3-d334-486f-a428-af1c8a3c360a/db032da3-d334-486f-a428-af1c8a3c360a.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 810.856731] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3597644b-f63d-4aaa-8079-0fc9227b27df {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.870025] env[63088]: DEBUG oslo_vmware.api [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Waiting for the task: (returnval){ [ 810.870025] env[63088]: value = "task-1284835" [ 810.870025] env[63088]: _type = "Task" [ 810.870025] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.875166] env[63088]: DEBUG oslo_vmware.api [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': task-1284835, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.916596] env[63088]: DEBUG nova.network.neutron [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 11ec987f-9908-4366-81ae-cbfe4de67b2d] Updating instance_info_cache with network_info: [{"id": "90ea719e-0a3d-4bae-a472-a58cbb3c1541", "address": "fa:16:3e:3c:08:40", "network": {"id": "5b801381-6bb3-45cd-b81f-92b5d757e080", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1481457494-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e48f62dee9ad4e9b94b67bd871db5c63", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "051f343d-ac4f-4070-a26d-467603122c81", "external-id": "nsx-vlan-transportzone-277", "segmentation_id": 277, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap90ea719e-0a", "ovs_interfaceid": "90ea719e-0a3d-4bae-a472-a58cbb3c1541", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 810.976036] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e29b656-4874-4c99-824f-1dbf00d9526f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.986463] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2cf36b6-07be-490a-b737-8c7c56618119 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.039335] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d082478-8212-4a2d-a7ac-5ad3e05dba84 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.051478] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-618fa2ad-8825-49c9-828b-b7607a3310e5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.072278] env[63088]: DEBUG nova.compute.provider_tree [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 811.378132] env[63088]: DEBUG oslo_vmware.api [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': task-1284835, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.456136} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.378132] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] db032da3-d334-486f-a428-af1c8a3c360a/db032da3-d334-486f-a428-af1c8a3c360a.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 811.378132] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: db032da3-d334-486f-a428-af1c8a3c360a] Extending root virtual disk to 1048576 {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 811.378132] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b8491d81-f711-4cc2-86b1-e68a305dd7a6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.384588] env[63088]: DEBUG oslo_vmware.api [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Waiting for the task: (returnval){ [ 811.384588] env[63088]: value = "task-1284836" [ 811.384588] env[63088]: _type = "Task" [ 811.384588] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.392299] env[63088]: DEBUG oslo_vmware.api [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': task-1284836, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.420318] env[63088]: DEBUG oslo_concurrency.lockutils [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Releasing lock "refresh_cache-11ec987f-9908-4366-81ae-cbfe4de67b2d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 811.421422] env[63088]: DEBUG nova.compute.manager [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 11ec987f-9908-4366-81ae-cbfe4de67b2d] Instance network_info: |[{"id": "90ea719e-0a3d-4bae-a472-a58cbb3c1541", "address": "fa:16:3e:3c:08:40", "network": {"id": "5b801381-6bb3-45cd-b81f-92b5d757e080", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1481457494-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e48f62dee9ad4e9b94b67bd871db5c63", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "051f343d-ac4f-4070-a26d-467603122c81", "external-id": "nsx-vlan-transportzone-277", "segmentation_id": 277, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap90ea719e-0a", "ovs_interfaceid": "90ea719e-0a3d-4bae-a472-a58cbb3c1541", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 811.421589] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 11ec987f-9908-4366-81ae-cbfe4de67b2d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3c:08:40', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '051f343d-ac4f-4070-a26d-467603122c81', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '90ea719e-0a3d-4bae-a472-a58cbb3c1541', 'vif_model': 'vmxnet3'}] {{(pid=63088) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 811.429154] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Creating folder: Project (e48f62dee9ad4e9b94b67bd871db5c63). Parent ref: group-v275816. {{(pid=63088) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 811.429439] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d087ba87-e1f7-40df-8e3d-7b45477d5852 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.441267] env[63088]: INFO nova.virt.vmwareapi.vm_util [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Created folder: Project (e48f62dee9ad4e9b94b67bd871db5c63) in parent group-v275816. [ 811.442018] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Creating folder: Instances. Parent ref: group-v275836. {{(pid=63088) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 811.442018] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3850a743-bd5a-4d94-a612-80d7e3e08618 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.451463] env[63088]: INFO nova.virt.vmwareapi.vm_util [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Created folder: Instances in parent group-v275836. [ 811.451686] env[63088]: DEBUG oslo.service.loopingcall [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 811.451956] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 11ec987f-9908-4366-81ae-cbfe4de67b2d] Creating VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 811.452110] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dc2ef6af-2180-46c2-9745-c0c19861aec7 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.470721] env[63088]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 811.470721] env[63088]: value = "task-1284839" [ 811.470721] env[63088]: _type = "Task" [ 811.470721] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.478283] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1284839, 'name': CreateVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.543656] env[63088]: DEBUG nova.compute.manager [req-59b81f06-432e-44df-a2e0-128e4295681a req-2e1ff743-b3c6-4c24-8958-b93bf73a0e9e service nova] [instance: 11ec987f-9908-4366-81ae-cbfe4de67b2d] Received event network-vif-plugged-90ea719e-0a3d-4bae-a472-a58cbb3c1541 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 811.543917] env[63088]: DEBUG oslo_concurrency.lockutils [req-59b81f06-432e-44df-a2e0-128e4295681a req-2e1ff743-b3c6-4c24-8958-b93bf73a0e9e service nova] Acquiring lock "11ec987f-9908-4366-81ae-cbfe4de67b2d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 811.544194] env[63088]: DEBUG oslo_concurrency.lockutils [req-59b81f06-432e-44df-a2e0-128e4295681a req-2e1ff743-b3c6-4c24-8958-b93bf73a0e9e service nova] Lock "11ec987f-9908-4366-81ae-cbfe4de67b2d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 811.544404] env[63088]: DEBUG oslo_concurrency.lockutils [req-59b81f06-432e-44df-a2e0-128e4295681a req-2e1ff743-b3c6-4c24-8958-b93bf73a0e9e service nova] Lock "11ec987f-9908-4366-81ae-cbfe4de67b2d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 811.544583] env[63088]: DEBUG nova.compute.manager [req-59b81f06-432e-44df-a2e0-128e4295681a req-2e1ff743-b3c6-4c24-8958-b93bf73a0e9e service nova] [instance: 11ec987f-9908-4366-81ae-cbfe4de67b2d] No waiting events found dispatching network-vif-plugged-90ea719e-0a3d-4bae-a472-a58cbb3c1541 {{(pid=63088) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 811.544747] env[63088]: WARNING nova.compute.manager [req-59b81f06-432e-44df-a2e0-128e4295681a req-2e1ff743-b3c6-4c24-8958-b93bf73a0e9e service nova] [instance: 11ec987f-9908-4366-81ae-cbfe4de67b2d] Received unexpected event network-vif-plugged-90ea719e-0a3d-4bae-a472-a58cbb3c1541 for instance with vm_state building and task_state spawning. [ 811.544924] env[63088]: DEBUG nova.compute.manager [req-59b81f06-432e-44df-a2e0-128e4295681a req-2e1ff743-b3c6-4c24-8958-b93bf73a0e9e service nova] [instance: 11ec987f-9908-4366-81ae-cbfe4de67b2d] Received event network-changed-90ea719e-0a3d-4bae-a472-a58cbb3c1541 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 811.545137] env[63088]: DEBUG nova.compute.manager [req-59b81f06-432e-44df-a2e0-128e4295681a req-2e1ff743-b3c6-4c24-8958-b93bf73a0e9e service nova] [instance: 11ec987f-9908-4366-81ae-cbfe4de67b2d] Refreshing instance network info cache due to event network-changed-90ea719e-0a3d-4bae-a472-a58cbb3c1541. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 811.545365] env[63088]: DEBUG oslo_concurrency.lockutils [req-59b81f06-432e-44df-a2e0-128e4295681a req-2e1ff743-b3c6-4c24-8958-b93bf73a0e9e service nova] Acquiring lock "refresh_cache-11ec987f-9908-4366-81ae-cbfe4de67b2d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 811.545531] env[63088]: DEBUG oslo_concurrency.lockutils [req-59b81f06-432e-44df-a2e0-128e4295681a req-2e1ff743-b3c6-4c24-8958-b93bf73a0e9e service nova] Acquired lock "refresh_cache-11ec987f-9908-4366-81ae-cbfe4de67b2d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 811.545695] env[63088]: DEBUG nova.network.neutron [req-59b81f06-432e-44df-a2e0-128e4295681a req-2e1ff743-b3c6-4c24-8958-b93bf73a0e9e service nova] [instance: 11ec987f-9908-4366-81ae-cbfe4de67b2d] Refreshing network info cache for port 90ea719e-0a3d-4bae-a472-a58cbb3c1541 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 811.575447] env[63088]: DEBUG nova.scheduler.client.report [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 811.708746] env[63088]: DEBUG nova.compute.manager [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 811.734964] env[63088]: DEBUG nova.virt.hardware [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 811.735228] env[63088]: DEBUG nova.virt.hardware [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 811.735385] env[63088]: DEBUG nova.virt.hardware [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 811.735565] env[63088]: DEBUG nova.virt.hardware [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 811.735712] env[63088]: DEBUG nova.virt.hardware [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 811.735859] env[63088]: DEBUG nova.virt.hardware [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 811.736070] env[63088]: DEBUG nova.virt.hardware [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 811.736229] env[63088]: DEBUG nova.virt.hardware [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 811.736391] env[63088]: DEBUG nova.virt.hardware [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 811.736550] env[63088]: DEBUG nova.virt.hardware [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 811.736720] env[63088]: DEBUG nova.virt.hardware [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 811.737630] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dac6cc6c-866b-4802-b79c-b37c2e2b6eb1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.745492] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b41490e7-e991-4c7c-8a7b-76ffaddd62c4 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.669039] env[63088]: DEBUG oslo_concurrency.lockutils [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.985s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 812.669494] env[63088]: DEBUG nova.compute.manager [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 812.672195] env[63088]: DEBUG oslo_vmware.api [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': task-1284836, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.060182} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.673257] env[63088]: DEBUG nova.network.neutron [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Successfully updated port: d5a10951-ae2e-48f5-b0ca-b1f144e5fe3a {{(pid=63088) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 812.677712] env[63088]: DEBUG nova.compute.manager [req-51da04ed-1af8-404d-89f2-5440104c7c55 req-7ded74e3-87e0-4023-bc13-7e5cd92dce6e service nova] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Received event network-vif-plugged-d5a10951-ae2e-48f5-b0ca-b1f144e5fe3a {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 812.678229] env[63088]: DEBUG oslo_concurrency.lockutils [req-51da04ed-1af8-404d-89f2-5440104c7c55 req-7ded74e3-87e0-4023-bc13-7e5cd92dce6e service nova] Acquiring lock "4971b24c-6710-4f50-9846-727dad264b1f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 812.678229] env[63088]: DEBUG oslo_concurrency.lockutils [req-51da04ed-1af8-404d-89f2-5440104c7c55 req-7ded74e3-87e0-4023-bc13-7e5cd92dce6e service nova] Lock "4971b24c-6710-4f50-9846-727dad264b1f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 812.678486] env[63088]: DEBUG oslo_concurrency.lockutils [req-51da04ed-1af8-404d-89f2-5440104c7c55 req-7ded74e3-87e0-4023-bc13-7e5cd92dce6e service nova] Lock "4971b24c-6710-4f50-9846-727dad264b1f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 812.678486] env[63088]: DEBUG nova.compute.manager [req-51da04ed-1af8-404d-89f2-5440104c7c55 req-7ded74e3-87e0-4023-bc13-7e5cd92dce6e service nova] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] No waiting events found dispatching network-vif-plugged-d5a10951-ae2e-48f5-b0ca-b1f144e5fe3a {{(pid=63088) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 812.678690] env[63088]: WARNING nova.compute.manager [req-51da04ed-1af8-404d-89f2-5440104c7c55 req-7ded74e3-87e0-4023-bc13-7e5cd92dce6e service nova] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Received unexpected event network-vif-plugged-d5a10951-ae2e-48f5-b0ca-b1f144e5fe3a for instance with vm_state building and task_state spawning. [ 812.678836] env[63088]: DEBUG oslo_concurrency.lockutils [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.871s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 812.680217] env[63088]: INFO nova.compute.claims [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: e845f60b-004f-4a5c-813e-0c183e99eba6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 812.682406] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: db032da3-d334-486f-a428-af1c8a3c360a] Extended root virtual disk {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 812.683775] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd43917c-76f0-45d3-876b-f070475a3f31 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.693167] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1284839, 'name': CreateVM_Task, 'duration_secs': 0.288544} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.702159] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 11ec987f-9908-4366-81ae-cbfe4de67b2d] Created VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 812.711284] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: db032da3-d334-486f-a428-af1c8a3c360a] Reconfiguring VM instance instance-00000034 to attach disk [datastore1] db032da3-d334-486f-a428-af1c8a3c360a/db032da3-d334-486f-a428-af1c8a3c360a.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 812.711918] env[63088]: DEBUG oslo_concurrency.lockutils [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 812.712104] env[63088]: DEBUG oslo_concurrency.lockutils [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 812.712407] env[63088]: DEBUG oslo_concurrency.lockutils [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 812.712890] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-18481404-2235-4294-b095-a1c2ac85a13d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.729406] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f0105a08-944e-4dc9-9483-6d9ff605ff29 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.735676] env[63088]: DEBUG oslo_vmware.api [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 812.735676] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52ce3f4e-b686-3754-5ded-9ca1d042f512" [ 812.735676] env[63088]: _type = "Task" [ 812.735676] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.740487] env[63088]: DEBUG oslo_vmware.api [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Waiting for the task: (returnval){ [ 812.740487] env[63088]: value = "task-1284840" [ 812.740487] env[63088]: _type = "Task" [ 812.740487] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.746868] env[63088]: DEBUG oslo_vmware.api [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52ce3f4e-b686-3754-5ded-9ca1d042f512, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.752087] env[63088]: DEBUG oslo_vmware.api [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': task-1284840, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.066965] env[63088]: DEBUG nova.network.neutron [req-59b81f06-432e-44df-a2e0-128e4295681a req-2e1ff743-b3c6-4c24-8958-b93bf73a0e9e service nova] [instance: 11ec987f-9908-4366-81ae-cbfe4de67b2d] Updated VIF entry in instance network info cache for port 90ea719e-0a3d-4bae-a472-a58cbb3c1541. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 813.067380] env[63088]: DEBUG nova.network.neutron [req-59b81f06-432e-44df-a2e0-128e4295681a req-2e1ff743-b3c6-4c24-8958-b93bf73a0e9e service nova] [instance: 11ec987f-9908-4366-81ae-cbfe4de67b2d] Updating instance_info_cache with network_info: [{"id": "90ea719e-0a3d-4bae-a472-a58cbb3c1541", "address": "fa:16:3e:3c:08:40", "network": {"id": "5b801381-6bb3-45cd-b81f-92b5d757e080", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1481457494-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e48f62dee9ad4e9b94b67bd871db5c63", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "051f343d-ac4f-4070-a26d-467603122c81", "external-id": "nsx-vlan-transportzone-277", "segmentation_id": 277, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap90ea719e-0a", "ovs_interfaceid": "90ea719e-0a3d-4bae-a472-a58cbb3c1541", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 813.174755] env[63088]: DEBUG nova.compute.utils [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 813.176135] env[63088]: DEBUG nova.compute.manager [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 813.176307] env[63088]: DEBUG nova.network.neutron [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 813.184258] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquiring lock "refresh_cache-4971b24c-6710-4f50-9846-727dad264b1f" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 813.184411] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquired lock "refresh_cache-4971b24c-6710-4f50-9846-727dad264b1f" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 813.185072] env[63088]: DEBUG nova.network.neutron [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 813.214905] env[63088]: DEBUG nova.policy [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4424adc5bbe34a91b669c3a907882644', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f33f2701fad94864a8c406a404bc0a42', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 813.254262] env[63088]: DEBUG oslo_vmware.api [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52ce3f4e-b686-3754-5ded-9ca1d042f512, 'name': SearchDatastore_Task, 'duration_secs': 0.009031} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.258496] env[63088]: DEBUG oslo_concurrency.lockutils [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 813.258800] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 11ec987f-9908-4366-81ae-cbfe4de67b2d] Processing image 1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 813.259084] env[63088]: DEBUG oslo_concurrency.lockutils [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 813.259201] env[63088]: DEBUG oslo_concurrency.lockutils [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 813.259376] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 813.259710] env[63088]: DEBUG oslo_vmware.api [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': task-1284840, 'name': ReconfigVM_Task, 'duration_secs': 0.259262} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.260225] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-41508c7a-9e4e-4875-9917-2f13c5d096e6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.261802] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: db032da3-d334-486f-a428-af1c8a3c360a] Reconfigured VM instance instance-00000034 to attach disk [datastore1] db032da3-d334-486f-a428-af1c8a3c360a/db032da3-d334-486f-a428-af1c8a3c360a.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 813.262697] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a59df553-ddf0-4298-8359-9da1b8e926d0 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.270999] env[63088]: DEBUG oslo_vmware.api [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Waiting for the task: (returnval){ [ 813.270999] env[63088]: value = "task-1284841" [ 813.270999] env[63088]: _type = "Task" [ 813.270999] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.272047] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 813.272260] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 813.275400] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9f15c745-5c3a-4caa-9f9c-c4f9caeb9f5f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.284082] env[63088]: DEBUG oslo_vmware.api [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': task-1284841, 'name': Rename_Task} progress is 6%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.285260] env[63088]: DEBUG oslo_vmware.api [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 813.285260] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]521872e5-fe9a-b60e-9487-baaa51474e70" [ 813.285260] env[63088]: _type = "Task" [ 813.285260] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.295304] env[63088]: DEBUG oslo_vmware.api [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]521872e5-fe9a-b60e-9487-baaa51474e70, 'name': SearchDatastore_Task, 'duration_secs': 0.008761} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.296568] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d339d249-57ab-4bbf-8c0f-f9ab9590d418 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.301432] env[63088]: DEBUG oslo_vmware.api [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 813.301432] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52933ac1-834e-aac8-3d1d-cf3cea06ab26" [ 813.301432] env[63088]: _type = "Task" [ 813.301432] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.308583] env[63088]: DEBUG oslo_vmware.api [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52933ac1-834e-aac8-3d1d-cf3cea06ab26, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.485611] env[63088]: DEBUG nova.network.neutron [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] Successfully created port: 53b4bd8d-df8b-4069-a75a-af46b605dddb {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 813.570733] env[63088]: DEBUG oslo_concurrency.lockutils [req-59b81f06-432e-44df-a2e0-128e4295681a req-2e1ff743-b3c6-4c24-8958-b93bf73a0e9e service nova] Releasing lock "refresh_cache-11ec987f-9908-4366-81ae-cbfe4de67b2d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 813.679971] env[63088]: DEBUG nova.compute.manager [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 813.738487] env[63088]: DEBUG nova.network.neutron [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 813.785316] env[63088]: DEBUG oslo_vmware.api [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': task-1284841, 'name': Rename_Task, 'duration_secs': 0.129119} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.787963] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: db032da3-d334-486f-a428-af1c8a3c360a] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 813.787963] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e6e859ac-b64c-4333-b3cd-ab3dbfa2972d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.798015] env[63088]: DEBUG oslo_vmware.api [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Waiting for the task: (returnval){ [ 813.798015] env[63088]: value = "task-1284842" [ 813.798015] env[63088]: _type = "Task" [ 813.798015] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.808470] env[63088]: DEBUG oslo_vmware.api [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': task-1284842, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.819986] env[63088]: DEBUG oslo_vmware.api [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52933ac1-834e-aac8-3d1d-cf3cea06ab26, 'name': SearchDatastore_Task, 'duration_secs': 0.008963} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.820753] env[63088]: DEBUG oslo_concurrency.lockutils [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 813.820753] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] 11ec987f-9908-4366-81ae-cbfe4de67b2d/11ec987f-9908-4366-81ae-cbfe4de67b2d.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 813.820900] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ce277739-a9ca-4269-b360-e45d4c162333 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.832418] env[63088]: DEBUG oslo_vmware.api [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 813.832418] env[63088]: value = "task-1284843" [ 813.832418] env[63088]: _type = "Task" [ 813.832418] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.843968] env[63088]: DEBUG oslo_vmware.api [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1284843, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.919569] env[63088]: DEBUG nova.network.neutron [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Updating instance_info_cache with network_info: [{"id": "d5a10951-ae2e-48f5-b0ca-b1f144e5fe3a", "address": "fa:16:3e:95:96:0c", "network": {"id": "1174a284-d3ca-4f9e-aa81-13ee9a693e55", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1994276040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa3d24a1a6c0430985fd80365d986ee1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd5a10951-ae", "ovs_interfaceid": "d5a10951-ae2e-48f5-b0ca-b1f144e5fe3a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 813.969450] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a88523d6-ff58-4e37-8067-b37b845a97dd {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.976907] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07373045-2083-41a3-ae2a-17d3b249425b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.008109] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f46ee3f2-f34e-4b74-bd19-427351e8038d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.015913] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b755524e-a8a8-48ec-bbc2-95dec2cefaab {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.029453] env[63088]: DEBUG nova.compute.provider_tree [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 814.060975] env[63088]: DEBUG nova.compute.manager [req-f1b97ea0-0722-489e-a042-084c2904063c req-a70bfc4d-c053-41f6-a6b2-c471d58567c2 service nova] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Received event network-changed-d5a10951-ae2e-48f5-b0ca-b1f144e5fe3a {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 814.061255] env[63088]: DEBUG nova.compute.manager [req-f1b97ea0-0722-489e-a042-084c2904063c req-a70bfc4d-c053-41f6-a6b2-c471d58567c2 service nova] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Refreshing instance network info cache due to event network-changed-d5a10951-ae2e-48f5-b0ca-b1f144e5fe3a. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 814.061387] env[63088]: DEBUG oslo_concurrency.lockutils [req-f1b97ea0-0722-489e-a042-084c2904063c req-a70bfc4d-c053-41f6-a6b2-c471d58567c2 service nova] Acquiring lock "refresh_cache-4971b24c-6710-4f50-9846-727dad264b1f" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 814.308206] env[63088]: DEBUG oslo_vmware.api [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': task-1284842, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.344789] env[63088]: DEBUG oslo_vmware.api [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1284843, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.46976} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.345021] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] 11ec987f-9908-4366-81ae-cbfe4de67b2d/11ec987f-9908-4366-81ae-cbfe4de67b2d.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 814.345291] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 11ec987f-9908-4366-81ae-cbfe4de67b2d] Extending root virtual disk to 1048576 {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 814.345582] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-19e94380-4e8f-4cdc-bcc2-e20fd1f6c3f6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.353524] env[63088]: DEBUG oslo_vmware.api [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 814.353524] env[63088]: value = "task-1284844" [ 814.353524] env[63088]: _type = "Task" [ 814.353524] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.362968] env[63088]: DEBUG oslo_vmware.api [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1284844, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.422673] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Releasing lock "refresh_cache-4971b24c-6710-4f50-9846-727dad264b1f" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 814.423093] env[63088]: DEBUG nova.compute.manager [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Instance network_info: |[{"id": "d5a10951-ae2e-48f5-b0ca-b1f144e5fe3a", "address": "fa:16:3e:95:96:0c", "network": {"id": "1174a284-d3ca-4f9e-aa81-13ee9a693e55", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1994276040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa3d24a1a6c0430985fd80365d986ee1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd5a10951-ae", "ovs_interfaceid": "d5a10951-ae2e-48f5-b0ca-b1f144e5fe3a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 814.423408] env[63088]: DEBUG oslo_concurrency.lockutils [req-f1b97ea0-0722-489e-a042-084c2904063c req-a70bfc4d-c053-41f6-a6b2-c471d58567c2 service nova] Acquired lock "refresh_cache-4971b24c-6710-4f50-9846-727dad264b1f" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 814.423602] env[63088]: DEBUG nova.network.neutron [req-f1b97ea0-0722-489e-a042-084c2904063c req-a70bfc4d-c053-41f6-a6b2-c471d58567c2 service nova] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Refreshing network info cache for port d5a10951-ae2e-48f5-b0ca-b1f144e5fe3a {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 814.425596] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:95:96:0c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd5a10951-ae2e-48f5-b0ca-b1f144e5fe3a', 'vif_model': 'vmxnet3'}] {{(pid=63088) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 814.433597] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Creating folder: Project (aa3d24a1a6c0430985fd80365d986ee1). Parent ref: group-v275816. {{(pid=63088) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 814.434611] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b23d6b20-e308-442d-ba5f-ce88733049ab {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.446389] env[63088]: INFO nova.virt.vmwareapi.vm_util [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Created folder: Project (aa3d24a1a6c0430985fd80365d986ee1) in parent group-v275816. [ 814.447104] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Creating folder: Instances. Parent ref: group-v275839. {{(pid=63088) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 814.447104] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-29a9ae88-8ab5-4b6c-990d-dccde2c97d99 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.455401] env[63088]: INFO nova.virt.vmwareapi.vm_util [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Created folder: Instances in parent group-v275839. [ 814.455625] env[63088]: DEBUG oslo.service.loopingcall [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 814.455808] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Creating VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 814.455996] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-34ff0d96-bf6f-4546-933f-93f0e4beacd8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.474443] env[63088]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 814.474443] env[63088]: value = "task-1284847" [ 814.474443] env[63088]: _type = "Task" [ 814.474443] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.481649] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1284847, 'name': CreateVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.532481] env[63088]: DEBUG nova.scheduler.client.report [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 814.690307] env[63088]: DEBUG nova.compute.manager [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 814.717407] env[63088]: DEBUG nova.virt.hardware [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 814.717653] env[63088]: DEBUG nova.virt.hardware [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 814.717810] env[63088]: DEBUG nova.virt.hardware [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 814.718011] env[63088]: DEBUG nova.virt.hardware [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 814.718170] env[63088]: DEBUG nova.virt.hardware [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 814.718320] env[63088]: DEBUG nova.virt.hardware [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 814.718530] env[63088]: DEBUG nova.virt.hardware [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 814.718690] env[63088]: DEBUG nova.virt.hardware [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 814.718857] env[63088]: DEBUG nova.virt.hardware [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 814.719028] env[63088]: DEBUG nova.virt.hardware [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 814.719208] env[63088]: DEBUG nova.virt.hardware [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 814.720631] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56a81842-1b59-4ab4-a0b2-2c1b452677f5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.727851] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-910aeebd-3862-4302-8d5d-ca38206b3f49 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.808069] env[63088]: DEBUG oslo_vmware.api [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': task-1284842, 'name': PowerOnVM_Task, 'duration_secs': 0.615542} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.808330] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: db032da3-d334-486f-a428-af1c8a3c360a] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 814.808528] env[63088]: INFO nova.compute.manager [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: db032da3-d334-486f-a428-af1c8a3c360a] Took 7.73 seconds to spawn the instance on the hypervisor. [ 814.808703] env[63088]: DEBUG nova.compute.manager [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: db032da3-d334-486f-a428-af1c8a3c360a] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 814.809483] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c8ff0cd-40bd-4940-a0e0-f08371504caa {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.864117] env[63088]: DEBUG oslo_vmware.api [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1284844, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069006} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.864429] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 11ec987f-9908-4366-81ae-cbfe4de67b2d] Extended root virtual disk {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 814.865194] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d609105-6ef9-4203-8760-32ba37eb7ed3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.888059] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 11ec987f-9908-4366-81ae-cbfe4de67b2d] Reconfiguring VM instance instance-00000035 to attach disk [datastore1] 11ec987f-9908-4366-81ae-cbfe4de67b2d/11ec987f-9908-4366-81ae-cbfe4de67b2d.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 814.888966] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-21cf862e-552a-4a9c-b879-7dddc57e70d6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.910988] env[63088]: DEBUG oslo_vmware.api [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 814.910988] env[63088]: value = "task-1284848" [ 814.910988] env[63088]: _type = "Task" [ 814.910988] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.917189] env[63088]: DEBUG oslo_vmware.api [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1284848, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.983924] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1284847, 'name': CreateVM_Task, 'duration_secs': 0.304359} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.984129] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Created VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 814.984860] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 814.985047] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 814.985399] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 814.986015] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5932cddf-49b6-41d5-87ac-302cc7514e88 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.992216] env[63088]: DEBUG oslo_vmware.api [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Waiting for the task: (returnval){ [ 814.992216] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52170aaa-611e-2864-2897-9441c0884e99" [ 814.992216] env[63088]: _type = "Task" [ 814.992216] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.000008] env[63088]: DEBUG oslo_vmware.api [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52170aaa-611e-2864-2897-9441c0884e99, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.037560] env[63088]: DEBUG oslo_concurrency.lockutils [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.359s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 815.038150] env[63088]: DEBUG nova.compute.manager [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: e845f60b-004f-4a5c-813e-0c183e99eba6] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 815.041168] env[63088]: DEBUG oslo_concurrency.lockutils [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.436s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 815.044809] env[63088]: INFO nova.compute.claims [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] [instance: 4f6e2d99-7300-42e6-8abf-362ddd4e8c6d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 815.053693] env[63088]: DEBUG nova.compute.manager [req-3eea15db-26a4-4dfb-a46d-e39f5bec5021 req-16591d05-47e6-4e6b-bdc1-90790c63b037 service nova] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] Received event network-vif-plugged-53b4bd8d-df8b-4069-a75a-af46b605dddb {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 815.053938] env[63088]: DEBUG oslo_concurrency.lockutils [req-3eea15db-26a4-4dfb-a46d-e39f5bec5021 req-16591d05-47e6-4e6b-bdc1-90790c63b037 service nova] Acquiring lock "4847f15e-98d4-401c-91ff-067e84a85727-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 815.054191] env[63088]: DEBUG oslo_concurrency.lockutils [req-3eea15db-26a4-4dfb-a46d-e39f5bec5021 req-16591d05-47e6-4e6b-bdc1-90790c63b037 service nova] Lock "4847f15e-98d4-401c-91ff-067e84a85727-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 815.054403] env[63088]: DEBUG oslo_concurrency.lockutils [req-3eea15db-26a4-4dfb-a46d-e39f5bec5021 req-16591d05-47e6-4e6b-bdc1-90790c63b037 service nova] Lock "4847f15e-98d4-401c-91ff-067e84a85727-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 815.054597] env[63088]: DEBUG nova.compute.manager [req-3eea15db-26a4-4dfb-a46d-e39f5bec5021 req-16591d05-47e6-4e6b-bdc1-90790c63b037 service nova] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] No waiting events found dispatching network-vif-plugged-53b4bd8d-df8b-4069-a75a-af46b605dddb {{(pid=63088) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 815.054805] env[63088]: WARNING nova.compute.manager [req-3eea15db-26a4-4dfb-a46d-e39f5bec5021 req-16591d05-47e6-4e6b-bdc1-90790c63b037 service nova] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] Received unexpected event network-vif-plugged-53b4bd8d-df8b-4069-a75a-af46b605dddb for instance with vm_state building and task_state spawning. [ 815.159291] env[63088]: DEBUG nova.network.neutron [req-f1b97ea0-0722-489e-a042-084c2904063c req-a70bfc4d-c053-41f6-a6b2-c471d58567c2 service nova] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Updated VIF entry in instance network info cache for port d5a10951-ae2e-48f5-b0ca-b1f144e5fe3a. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 815.159782] env[63088]: DEBUG nova.network.neutron [req-f1b97ea0-0722-489e-a042-084c2904063c req-a70bfc4d-c053-41f6-a6b2-c471d58567c2 service nova] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Updating instance_info_cache with network_info: [{"id": "d5a10951-ae2e-48f5-b0ca-b1f144e5fe3a", "address": "fa:16:3e:95:96:0c", "network": {"id": "1174a284-d3ca-4f9e-aa81-13ee9a693e55", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1994276040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa3d24a1a6c0430985fd80365d986ee1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd5a10951-ae", "ovs_interfaceid": "d5a10951-ae2e-48f5-b0ca-b1f144e5fe3a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 815.162412] env[63088]: DEBUG nova.network.neutron [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] Successfully updated port: 53b4bd8d-df8b-4069-a75a-af46b605dddb {{(pid=63088) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 815.325059] env[63088]: INFO nova.compute.manager [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: db032da3-d334-486f-a428-af1c8a3c360a] Took 35.00 seconds to build instance. [ 815.418929] env[63088]: DEBUG oslo_vmware.api [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1284848, 'name': ReconfigVM_Task, 'duration_secs': 0.267048} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.419697] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 11ec987f-9908-4366-81ae-cbfe4de67b2d] Reconfigured VM instance instance-00000035 to attach disk [datastore1] 11ec987f-9908-4366-81ae-cbfe4de67b2d/11ec987f-9908-4366-81ae-cbfe4de67b2d.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 815.420091] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-61b31fba-6bde-4bef-8a07-c9f375707e6a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.427026] env[63088]: DEBUG oslo_vmware.api [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 815.427026] env[63088]: value = "task-1284849" [ 815.427026] env[63088]: _type = "Task" [ 815.427026] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.435934] env[63088]: DEBUG oslo_vmware.api [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1284849, 'name': Rename_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.502369] env[63088]: DEBUG oslo_vmware.api [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52170aaa-611e-2864-2897-9441c0884e99, 'name': SearchDatastore_Task, 'duration_secs': 0.009366} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.502369] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 815.502369] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Processing image 1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 815.503257] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 815.503436] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 815.503637] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 815.503909] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-15b665f9-8755-4b83-adba-da443316771e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.512193] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 815.512380] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 815.513089] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1c1f53ca-993a-4e60-83c0-5f706dabe53f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.518054] env[63088]: DEBUG oslo_vmware.api [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Waiting for the task: (returnval){ [ 815.518054] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52a5c2da-204e-41dd-59e6-c65c40b419e0" [ 815.518054] env[63088]: _type = "Task" [ 815.518054] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.525649] env[63088]: DEBUG oslo_vmware.api [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52a5c2da-204e-41dd-59e6-c65c40b419e0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.547273] env[63088]: DEBUG nova.compute.utils [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 815.552525] env[63088]: DEBUG nova.compute.manager [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: e845f60b-004f-4a5c-813e-0c183e99eba6] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 815.552712] env[63088]: DEBUG nova.network.neutron [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: e845f60b-004f-4a5c-813e-0c183e99eba6] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 815.592112] env[63088]: DEBUG nova.policy [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '32d5ff9b3ea345e4ac56010aa917e51b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '03e20d606e654362acbe2b36fe499ae3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 815.664710] env[63088]: DEBUG oslo_concurrency.lockutils [req-f1b97ea0-0722-489e-a042-084c2904063c req-a70bfc4d-c053-41f6-a6b2-c471d58567c2 service nova] Releasing lock "refresh_cache-4971b24c-6710-4f50-9846-727dad264b1f" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 815.665355] env[63088]: DEBUG oslo_concurrency.lockutils [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquiring lock "refresh_cache-4847f15e-98d4-401c-91ff-067e84a85727" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 815.665481] env[63088]: DEBUG oslo_concurrency.lockutils [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquired lock "refresh_cache-4847f15e-98d4-401c-91ff-067e84a85727" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 815.665627] env[63088]: DEBUG nova.network.neutron [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 815.826526] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6ea1d6e5-6921-480e-9fed-30f7184c4776 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Lock "db032da3-d334-486f-a428-af1c8a3c360a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 120.489s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 815.855904] env[63088]: DEBUG nova.network.neutron [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: e845f60b-004f-4a5c-813e-0c183e99eba6] Successfully created port: 8ca1e82d-550d-4faa-8b7e-93e0b812c234 {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 815.942276] env[63088]: DEBUG oslo_vmware.api [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1284849, 'name': Rename_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.029171] env[63088]: DEBUG oslo_vmware.api [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52a5c2da-204e-41dd-59e6-c65c40b419e0, 'name': SearchDatastore_Task, 'duration_secs': 0.007854} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.033016] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f6f8a834-610e-4e84-8a56-2a1005bbdc97 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.035997] env[63088]: DEBUG oslo_vmware.api [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Waiting for the task: (returnval){ [ 816.035997] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52da1cf9-7a1c-53c5-980c-63b227e112ed" [ 816.035997] env[63088]: _type = "Task" [ 816.035997] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.044076] env[63088]: DEBUG oslo_vmware.api [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52da1cf9-7a1c-53c5-980c-63b227e112ed, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.055059] env[63088]: DEBUG nova.compute.manager [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: e845f60b-004f-4a5c-813e-0c183e99eba6] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 816.218242] env[63088]: DEBUG nova.network.neutron [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 816.328645] env[63088]: DEBUG nova.compute.manager [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] [instance: 1d421bf0-925d-408c-a612-b659da92799f] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 816.335288] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a29e989-a77a-46e3-9ded-1142d5d58ac4 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.342848] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6034b8dd-7ab7-435a-aa8a-067b395fe7f3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.376238] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e68005f-0c0a-4d75-bb50-a1c788720827 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.381294] env[63088]: DEBUG nova.network.neutron [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] Updating instance_info_cache with network_info: [{"id": "53b4bd8d-df8b-4069-a75a-af46b605dddb", "address": "fa:16:3e:6d:a9:02", "network": {"id": "dff14a7f-0af1-4e4e-a498-86d7c9816e6b", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-762073671-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f33f2701fad94864a8c406a404bc0a42", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "21310d90-efbc-45a8-a97f-c4358606530f", "external-id": "nsx-vlan-transportzone-672", "segmentation_id": 672, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap53b4bd8d-df", "ovs_interfaceid": "53b4bd8d-df8b-4069-a75a-af46b605dddb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 816.386142] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea63d199-cc22-4364-b66c-e9ec9b003496 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.400413] env[63088]: DEBUG nova.compute.provider_tree [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 816.436036] env[63088]: INFO nova.compute.manager [None req-69612416-d1eb-4a97-8ff2-4c55509ccebf tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: db032da3-d334-486f-a428-af1c8a3c360a] Rescuing [ 816.436845] env[63088]: DEBUG oslo_concurrency.lockutils [None req-69612416-d1eb-4a97-8ff2-4c55509ccebf tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Acquiring lock "refresh_cache-db032da3-d334-486f-a428-af1c8a3c360a" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 816.436845] env[63088]: DEBUG oslo_concurrency.lockutils [None req-69612416-d1eb-4a97-8ff2-4c55509ccebf tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Acquired lock "refresh_cache-db032da3-d334-486f-a428-af1c8a3c360a" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.436845] env[63088]: DEBUG nova.network.neutron [None req-69612416-d1eb-4a97-8ff2-4c55509ccebf tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: db032da3-d334-486f-a428-af1c8a3c360a] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 816.440601] env[63088]: DEBUG oslo_vmware.api [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1284849, 'name': Rename_Task, 'duration_secs': 0.818136} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.441032] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 11ec987f-9908-4366-81ae-cbfe4de67b2d] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 816.441262] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e6a69fb6-1a36-484a-ae0b-5ad52488f3c2 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.447921] env[63088]: DEBUG oslo_vmware.api [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 816.447921] env[63088]: value = "task-1284850" [ 816.447921] env[63088]: _type = "Task" [ 816.447921] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.455265] env[63088]: DEBUG oslo_vmware.api [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1284850, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.546472] env[63088]: DEBUG oslo_vmware.api [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52da1cf9-7a1c-53c5-980c-63b227e112ed, 'name': SearchDatastore_Task, 'duration_secs': 0.008432} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.546769] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 816.547068] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] 4971b24c-6710-4f50-9846-727dad264b1f/4971b24c-6710-4f50-9846-727dad264b1f.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 816.547342] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5fb69f61-a4c9-4a58-903b-53be437a314a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.553629] env[63088]: DEBUG oslo_vmware.api [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Waiting for the task: (returnval){ [ 816.553629] env[63088]: value = "task-1284851" [ 816.553629] env[63088]: _type = "Task" [ 816.553629] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.564189] env[63088]: DEBUG oslo_vmware.api [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1284851, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.850042] env[63088]: DEBUG oslo_concurrency.lockutils [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 816.885422] env[63088]: DEBUG oslo_concurrency.lockutils [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Releasing lock "refresh_cache-4847f15e-98d4-401c-91ff-067e84a85727" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 816.885774] env[63088]: DEBUG nova.compute.manager [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] Instance network_info: |[{"id": "53b4bd8d-df8b-4069-a75a-af46b605dddb", "address": "fa:16:3e:6d:a9:02", "network": {"id": "dff14a7f-0af1-4e4e-a498-86d7c9816e6b", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-762073671-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f33f2701fad94864a8c406a404bc0a42", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "21310d90-efbc-45a8-a97f-c4358606530f", "external-id": "nsx-vlan-transportzone-672", "segmentation_id": 672, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap53b4bd8d-df", "ovs_interfaceid": "53b4bd8d-df8b-4069-a75a-af46b605dddb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 816.886300] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6d:a9:02', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '21310d90-efbc-45a8-a97f-c4358606530f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '53b4bd8d-df8b-4069-a75a-af46b605dddb', 'vif_model': 'vmxnet3'}] {{(pid=63088) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 816.896062] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Creating folder: Project (f33f2701fad94864a8c406a404bc0a42). Parent ref: group-v275816. {{(pid=63088) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 816.896434] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2a816812-7d45-4dac-adaf-1e9d7e10fb70 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.902812] env[63088]: DEBUG nova.scheduler.client.report [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 816.908289] env[63088]: INFO nova.virt.vmwareapi.vm_util [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Created folder: Project (f33f2701fad94864a8c406a404bc0a42) in parent group-v275816. [ 816.908494] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Creating folder: Instances. Parent ref: group-v275842. {{(pid=63088) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 816.909011] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2a957ba6-29d6-4efe-a468-abe10dee263e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.919351] env[63088]: INFO nova.virt.vmwareapi.vm_util [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Created folder: Instances in parent group-v275842. [ 816.919693] env[63088]: DEBUG oslo.service.loopingcall [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 816.920786] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] Creating VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 816.921072] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b130cbda-ed1c-4545-9530-b7197b030c83 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.944947] env[63088]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 816.944947] env[63088]: value = "task-1284854" [ 816.944947] env[63088]: _type = "Task" [ 816.944947] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.962541] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1284854, 'name': CreateVM_Task} progress is 6%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.966339] env[63088]: DEBUG oslo_vmware.api [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1284850, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.068924] env[63088]: DEBUG nova.compute.manager [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: e845f60b-004f-4a5c-813e-0c183e99eba6] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 817.071100] env[63088]: DEBUG oslo_vmware.api [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1284851, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.50208} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.071692] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] 4971b24c-6710-4f50-9846-727dad264b1f/4971b24c-6710-4f50-9846-727dad264b1f.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 817.071789] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Extending root virtual disk to 1048576 {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 817.072331] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c423d04b-c01f-4aea-bec7-a72a8401ecaa {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.079232] env[63088]: DEBUG oslo_vmware.api [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Waiting for the task: (returnval){ [ 817.079232] env[63088]: value = "task-1284855" [ 817.079232] env[63088]: _type = "Task" [ 817.079232] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.087726] env[63088]: DEBUG oslo_vmware.api [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1284855, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.101089] env[63088]: DEBUG nova.virt.hardware [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 817.101405] env[63088]: DEBUG nova.virt.hardware [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 817.101602] env[63088]: DEBUG nova.virt.hardware [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 817.101821] env[63088]: DEBUG nova.virt.hardware [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 817.101997] env[63088]: DEBUG nova.virt.hardware [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 817.102232] env[63088]: DEBUG nova.virt.hardware [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 817.102567] env[63088]: DEBUG nova.virt.hardware [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 817.102812] env[63088]: DEBUG nova.virt.hardware [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 817.103068] env[63088]: DEBUG nova.virt.hardware [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 817.103301] env[63088]: DEBUG nova.virt.hardware [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 817.103537] env[63088]: DEBUG nova.virt.hardware [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 817.104574] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90458688-0793-45f8-a021-ac904dc4e3cb {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.113028] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73cad922-1b22-45b0-96d7-946a2e57a6cb {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.154271] env[63088]: DEBUG nova.compute.manager [req-6f1ee4f8-b110-478f-b9b5-4a1252ebc031 req-36347efe-e213-4838-a50c-3e5f261ee545 service nova] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] Received event network-changed-53b4bd8d-df8b-4069-a75a-af46b605dddb {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 817.154354] env[63088]: DEBUG nova.compute.manager [req-6f1ee4f8-b110-478f-b9b5-4a1252ebc031 req-36347efe-e213-4838-a50c-3e5f261ee545 service nova] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] Refreshing instance network info cache due to event network-changed-53b4bd8d-df8b-4069-a75a-af46b605dddb. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 817.154613] env[63088]: DEBUG oslo_concurrency.lockutils [req-6f1ee4f8-b110-478f-b9b5-4a1252ebc031 req-36347efe-e213-4838-a50c-3e5f261ee545 service nova] Acquiring lock "refresh_cache-4847f15e-98d4-401c-91ff-067e84a85727" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 817.155015] env[63088]: DEBUG oslo_concurrency.lockutils [req-6f1ee4f8-b110-478f-b9b5-4a1252ebc031 req-36347efe-e213-4838-a50c-3e5f261ee545 service nova] Acquired lock "refresh_cache-4847f15e-98d4-401c-91ff-067e84a85727" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 817.155015] env[63088]: DEBUG nova.network.neutron [req-6f1ee4f8-b110-478f-b9b5-4a1252ebc031 req-36347efe-e213-4838-a50c-3e5f261ee545 service nova] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] Refreshing network info cache for port 53b4bd8d-df8b-4069-a75a-af46b605dddb {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 817.220372] env[63088]: DEBUG nova.network.neutron [None req-69612416-d1eb-4a97-8ff2-4c55509ccebf tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: db032da3-d334-486f-a428-af1c8a3c360a] Updating instance_info_cache with network_info: [{"id": "98219984-3198-4e54-8c93-024446f958a3", "address": "fa:16:3e:6d:e6:fe", "network": {"id": "7db9302e-1c7a-42dc-8084-98e2a57ad7bf", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-2073161169-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "900466c3c09a4f928bdd4b602715c838", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "87bbf4e0-9064-4516-b7e7-44973f817205", "external-id": "nsx-vlan-transportzone-507", "segmentation_id": 507, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap98219984-31", "ovs_interfaceid": "98219984-3198-4e54-8c93-024446f958a3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 817.410070] env[63088]: DEBUG oslo_concurrency.lockutils [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.369s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 817.410609] env[63088]: DEBUG nova.compute.manager [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] [instance: 4f6e2d99-7300-42e6-8abf-362ddd4e8c6d] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 817.413393] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.932s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 817.417019] env[63088]: INFO nova.compute.claims [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 817.458046] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1284854, 'name': CreateVM_Task, 'duration_secs': 0.393608} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.458372] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] Created VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 817.459094] env[63088]: DEBUG oslo_concurrency.lockutils [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 817.459266] env[63088]: DEBUG oslo_concurrency.lockutils [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 817.459603] env[63088]: DEBUG oslo_concurrency.lockutils [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 817.459853] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ebceb89c-825a-42fc-bb17-836e22123e13 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.465473] env[63088]: DEBUG oslo_vmware.api [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1284850, 'name': PowerOnVM_Task, 'duration_secs': 0.825836} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.466039] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 11ec987f-9908-4366-81ae-cbfe4de67b2d] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 817.466272] env[63088]: INFO nova.compute.manager [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 11ec987f-9908-4366-81ae-cbfe4de67b2d] Took 8.06 seconds to spawn the instance on the hypervisor. [ 817.466461] env[63088]: DEBUG nova.compute.manager [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 11ec987f-9908-4366-81ae-cbfe4de67b2d] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 817.467328] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a21d9e9f-641b-4e71-b698-f05cad744f10 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.471070] env[63088]: DEBUG oslo_vmware.api [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Waiting for the task: (returnval){ [ 817.471070] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52e61fe6-6037-f2f5-3bef-8f1eed9eb366" [ 817.471070] env[63088]: _type = "Task" [ 817.471070] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.482742] env[63088]: DEBUG oslo_vmware.api [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52e61fe6-6037-f2f5-3bef-8f1eed9eb366, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.499347] env[63088]: DEBUG nova.network.neutron [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: e845f60b-004f-4a5c-813e-0c183e99eba6] Successfully updated port: 8ca1e82d-550d-4faa-8b7e-93e0b812c234 {{(pid=63088) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 817.589531] env[63088]: DEBUG oslo_vmware.api [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1284855, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072454} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.589945] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Extended root virtual disk {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 817.590827] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5a56ea4-30b9-4aef-978e-17dc4fd7d321 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.613465] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Reconfiguring VM instance instance-00000036 to attach disk [datastore1] 4971b24c-6710-4f50-9846-727dad264b1f/4971b24c-6710-4f50-9846-727dad264b1f.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 817.613800] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1724faa2-08b4-40a8-8efa-93efdad3dc91 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.633398] env[63088]: DEBUG oslo_vmware.api [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Waiting for the task: (returnval){ [ 817.633398] env[63088]: value = "task-1284856" [ 817.633398] env[63088]: _type = "Task" [ 817.633398] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.641828] env[63088]: DEBUG oslo_vmware.api [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1284856, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.723427] env[63088]: DEBUG oslo_concurrency.lockutils [None req-69612416-d1eb-4a97-8ff2-4c55509ccebf tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Releasing lock "refresh_cache-db032da3-d334-486f-a428-af1c8a3c360a" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 817.883342] env[63088]: DEBUG nova.network.neutron [req-6f1ee4f8-b110-478f-b9b5-4a1252ebc031 req-36347efe-e213-4838-a50c-3e5f261ee545 service nova] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] Updated VIF entry in instance network info cache for port 53b4bd8d-df8b-4069-a75a-af46b605dddb. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 817.883714] env[63088]: DEBUG nova.network.neutron [req-6f1ee4f8-b110-478f-b9b5-4a1252ebc031 req-36347efe-e213-4838-a50c-3e5f261ee545 service nova] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] Updating instance_info_cache with network_info: [{"id": "53b4bd8d-df8b-4069-a75a-af46b605dddb", "address": "fa:16:3e:6d:a9:02", "network": {"id": "dff14a7f-0af1-4e4e-a498-86d7c9816e6b", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-762073671-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f33f2701fad94864a8c406a404bc0a42", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "21310d90-efbc-45a8-a97f-c4358606530f", "external-id": "nsx-vlan-transportzone-672", "segmentation_id": 672, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap53b4bd8d-df", "ovs_interfaceid": "53b4bd8d-df8b-4069-a75a-af46b605dddb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 817.923077] env[63088]: DEBUG nova.compute.utils [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 817.926095] env[63088]: DEBUG nova.compute.manager [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] [instance: 4f6e2d99-7300-42e6-8abf-362ddd4e8c6d] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 817.926286] env[63088]: DEBUG nova.network.neutron [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] [instance: 4f6e2d99-7300-42e6-8abf-362ddd4e8c6d] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 817.965647] env[63088]: DEBUG nova.policy [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5ab31869a6324502aac2689b2d675e8f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '960b3ebce4b747dfa84c1659118c78f0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 817.984908] env[63088]: DEBUG oslo_vmware.api [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52e61fe6-6037-f2f5-3bef-8f1eed9eb366, 'name': SearchDatastore_Task, 'duration_secs': 0.043561} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.987500] env[63088]: DEBUG oslo_concurrency.lockutils [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 817.987874] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] Processing image 1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 817.987947] env[63088]: DEBUG oslo_concurrency.lockutils [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 817.988089] env[63088]: DEBUG oslo_concurrency.lockutils [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 817.988269] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 817.988739] env[63088]: INFO nova.compute.manager [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 11ec987f-9908-4366-81ae-cbfe4de67b2d] Took 35.86 seconds to build instance. [ 817.989617] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ab218bc1-515d-468e-a2e5-fc30642c6b53 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.997454] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 817.997661] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 817.998386] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-50124695-c84b-4ab7-ac76-a826b1728746 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.001966] env[63088]: DEBUG oslo_concurrency.lockutils [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquiring lock "refresh_cache-e845f60b-004f-4a5c-813e-0c183e99eba6" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 818.001966] env[63088]: DEBUG oslo_concurrency.lockutils [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquired lock "refresh_cache-e845f60b-004f-4a5c-813e-0c183e99eba6" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 818.001966] env[63088]: DEBUG nova.network.neutron [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: e845f60b-004f-4a5c-813e-0c183e99eba6] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 818.006478] env[63088]: DEBUG oslo_vmware.api [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Waiting for the task: (returnval){ [ 818.006478] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]521b4cf4-edb4-8c17-9305-cd583e4c3b54" [ 818.006478] env[63088]: _type = "Task" [ 818.006478] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.019053] env[63088]: DEBUG oslo_vmware.api [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]521b4cf4-edb4-8c17-9305-cd583e4c3b54, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.142979] env[63088]: DEBUG oslo_vmware.api [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1284856, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.237407] env[63088]: DEBUG nova.network.neutron [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] [instance: 4f6e2d99-7300-42e6-8abf-362ddd4e8c6d] Successfully created port: 07c78e97-0b3c-45a7-859f-bbf74ce3c5f6 {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 818.250409] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-69612416-d1eb-4a97-8ff2-4c55509ccebf tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: db032da3-d334-486f-a428-af1c8a3c360a] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 818.250751] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-97c9d881-4f01-4aff-8238-97b433eceed7 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.258938] env[63088]: DEBUG oslo_vmware.api [None req-69612416-d1eb-4a97-8ff2-4c55509ccebf tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Waiting for the task: (returnval){ [ 818.258938] env[63088]: value = "task-1284857" [ 818.258938] env[63088]: _type = "Task" [ 818.258938] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.267567] env[63088]: DEBUG oslo_vmware.api [None req-69612416-d1eb-4a97-8ff2-4c55509ccebf tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': task-1284857, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.386809] env[63088]: DEBUG oslo_concurrency.lockutils [req-6f1ee4f8-b110-478f-b9b5-4a1252ebc031 req-36347efe-e213-4838-a50c-3e5f261ee545 service nova] Releasing lock "refresh_cache-4847f15e-98d4-401c-91ff-067e84a85727" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 818.427547] env[63088]: DEBUG nova.compute.manager [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] [instance: 4f6e2d99-7300-42e6-8abf-362ddd4e8c6d] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 818.492527] env[63088]: DEBUG oslo_concurrency.lockutils [None req-9de6a4d9-fd7a-4942-963a-bfa9911a2530 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "11ec987f-9908-4366-81ae-cbfe4de67b2d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 101.502s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 818.524318] env[63088]: DEBUG oslo_vmware.api [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]521b4cf4-edb4-8c17-9305-cd583e4c3b54, 'name': SearchDatastore_Task, 'duration_secs': 0.03079} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.525253] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-19373f4e-caa6-41c9-b930-040090e8898e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.534079] env[63088]: DEBUG oslo_vmware.api [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Waiting for the task: (returnval){ [ 818.534079] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52d767ae-c09c-a465-c3b2-8ac43312bc13" [ 818.534079] env[63088]: _type = "Task" [ 818.534079] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.542442] env[63088]: DEBUG oslo_vmware.api [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52d767ae-c09c-a465-c3b2-8ac43312bc13, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.552335] env[63088]: DEBUG nova.network.neutron [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: e845f60b-004f-4a5c-813e-0c183e99eba6] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 818.645736] env[63088]: DEBUG oslo_vmware.api [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1284856, 'name': ReconfigVM_Task, 'duration_secs': 0.929767} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.645987] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Reconfigured VM instance instance-00000036 to attach disk [datastore1] 4971b24c-6710-4f50-9846-727dad264b1f/4971b24c-6710-4f50-9846-727dad264b1f.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 818.648678] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e432fba1-34ef-4c73-99c1-e1f977b6dca2 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.655539] env[63088]: DEBUG oslo_vmware.api [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Waiting for the task: (returnval){ [ 818.655539] env[63088]: value = "task-1284858" [ 818.655539] env[63088]: _type = "Task" [ 818.655539] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.664983] env[63088]: DEBUG oslo_vmware.api [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1284858, 'name': Rename_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.721224] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ac00b11-8176-42d7-bc3d-411a96c5a5ef {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.728711] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f9b9c7b-9586-4873-93b7-d0c53fa8e448 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.761998] env[63088]: DEBUG nova.network.neutron [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: e845f60b-004f-4a5c-813e-0c183e99eba6] Updating instance_info_cache with network_info: [{"id": "8ca1e82d-550d-4faa-8b7e-93e0b812c234", "address": "fa:16:3e:11:39:ef", "network": {"id": "307b966a-d9e1-40fd-9313-1ad94c734308", "bridge": "br-int", "label": "tempest-ServersTestJSON-1907249197-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "03e20d606e654362acbe2b36fe499ae3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae18b41f-e73c-44f1-83dd-467c080944f4", "external-id": "nsx-vlan-transportzone-653", "segmentation_id": 653, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8ca1e82d-55", "ovs_interfaceid": "8ca1e82d-550d-4faa-8b7e-93e0b812c234", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 818.766466] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ade3e4f-339d-490f-ae7d-c6c6b814dbc5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.778684] env[63088]: DEBUG oslo_vmware.api [None req-69612416-d1eb-4a97-8ff2-4c55509ccebf tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': task-1284857, 'name': PowerOffVM_Task, 'duration_secs': 0.30378} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.779036] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-69612416-d1eb-4a97-8ff2-4c55509ccebf tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: db032da3-d334-486f-a428-af1c8a3c360a] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 818.780292] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fec7ec18-3a6c-4e93-b5af-cffbf25ebcbb {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.785884] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb4e8444-121b-42cb-ba78-ca7e46b98678 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.798961] env[63088]: DEBUG nova.compute.provider_tree [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 818.813625] env[63088]: DEBUG nova.scheduler.client.report [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 818.820016] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c60aac8-fe30-40d1-831c-2d167628dcf4 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.891158] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-69612416-d1eb-4a97-8ff2-4c55509ccebf tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: db032da3-d334-486f-a428-af1c8a3c360a] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 818.891857] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3a514dcf-4564-4fcf-a95d-a6609cf1b25e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.900158] env[63088]: DEBUG oslo_vmware.api [None req-69612416-d1eb-4a97-8ff2-4c55509ccebf tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Waiting for the task: (returnval){ [ 818.900158] env[63088]: value = "task-1284859" [ 818.900158] env[63088]: _type = "Task" [ 818.900158] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.908425] env[63088]: DEBUG oslo_vmware.api [None req-69612416-d1eb-4a97-8ff2-4c55509ccebf tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': task-1284859, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.995252] env[63088]: DEBUG nova.compute.manager [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] [instance: 2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 819.044528] env[63088]: DEBUG oslo_vmware.api [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52d767ae-c09c-a465-c3b2-8ac43312bc13, 'name': SearchDatastore_Task, 'duration_secs': 0.015607} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.044820] env[63088]: DEBUG oslo_concurrency.lockutils [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 819.045095] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore2] 4847f15e-98d4-401c-91ff-067e84a85727/4847f15e-98d4-401c-91ff-067e84a85727.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 819.045352] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-24aae3ae-6bc7-4f39-800a-4e86711613eb {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.051936] env[63088]: DEBUG oslo_vmware.api [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Waiting for the task: (returnval){ [ 819.051936] env[63088]: value = "task-1284860" [ 819.051936] env[63088]: _type = "Task" [ 819.051936] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.059288] env[63088]: DEBUG oslo_vmware.api [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1284860, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.128795] env[63088]: DEBUG oslo_concurrency.lockutils [None req-da2e38c8-9951-4dcc-950e-b6d060a5ab7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquiring lock "11ec987f-9908-4366-81ae-cbfe4de67b2d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 819.129075] env[63088]: DEBUG oslo_concurrency.lockutils [None req-da2e38c8-9951-4dcc-950e-b6d060a5ab7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "11ec987f-9908-4366-81ae-cbfe4de67b2d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 819.129287] env[63088]: DEBUG oslo_concurrency.lockutils [None req-da2e38c8-9951-4dcc-950e-b6d060a5ab7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquiring lock "11ec987f-9908-4366-81ae-cbfe4de67b2d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 819.129466] env[63088]: DEBUG oslo_concurrency.lockutils [None req-da2e38c8-9951-4dcc-950e-b6d060a5ab7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "11ec987f-9908-4366-81ae-cbfe4de67b2d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 819.129656] env[63088]: DEBUG oslo_concurrency.lockutils [None req-da2e38c8-9951-4dcc-950e-b6d060a5ab7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "11ec987f-9908-4366-81ae-cbfe4de67b2d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 819.131715] env[63088]: INFO nova.compute.manager [None req-da2e38c8-9951-4dcc-950e-b6d060a5ab7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 11ec987f-9908-4366-81ae-cbfe4de67b2d] Terminating instance [ 819.133442] env[63088]: DEBUG nova.compute.manager [None req-da2e38c8-9951-4dcc-950e-b6d060a5ab7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 11ec987f-9908-4366-81ae-cbfe4de67b2d] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 819.133631] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-da2e38c8-9951-4dcc-950e-b6d060a5ab7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 11ec987f-9908-4366-81ae-cbfe4de67b2d] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 819.134446] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c68bf27-c3a8-4b13-9331-4b912616303b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.142255] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-da2e38c8-9951-4dcc-950e-b6d060a5ab7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 11ec987f-9908-4366-81ae-cbfe4de67b2d] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 819.142477] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bf803512-162d-4a74-a445-283b44881584 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.148358] env[63088]: DEBUG oslo_vmware.api [None req-da2e38c8-9951-4dcc-950e-b6d060a5ab7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 819.148358] env[63088]: value = "task-1284861" [ 819.148358] env[63088]: _type = "Task" [ 819.148358] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.155756] env[63088]: DEBUG oslo_vmware.api [None req-da2e38c8-9951-4dcc-950e-b6d060a5ab7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1284861, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.165066] env[63088]: DEBUG oslo_vmware.api [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1284858, 'name': Rename_Task, 'duration_secs': 0.272103} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.165365] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 819.165594] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e133fe9c-f9f6-4bfa-a20e-9ec8f5e6c558 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.171976] env[63088]: DEBUG oslo_vmware.api [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Waiting for the task: (returnval){ [ 819.171976] env[63088]: value = "task-1284862" [ 819.171976] env[63088]: _type = "Task" [ 819.171976] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.179469] env[63088]: DEBUG oslo_vmware.api [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1284862, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.182110] env[63088]: DEBUG nova.compute.manager [req-2e64b50e-38f8-42ea-b6be-00cc1da8c766 req-4ca5e16e-c313-495d-b2c2-46426958694f service nova] [instance: e845f60b-004f-4a5c-813e-0c183e99eba6] Received event network-vif-plugged-8ca1e82d-550d-4faa-8b7e-93e0b812c234 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 819.182110] env[63088]: DEBUG oslo_concurrency.lockutils [req-2e64b50e-38f8-42ea-b6be-00cc1da8c766 req-4ca5e16e-c313-495d-b2c2-46426958694f service nova] Acquiring lock "e845f60b-004f-4a5c-813e-0c183e99eba6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 819.182110] env[63088]: DEBUG oslo_concurrency.lockutils [req-2e64b50e-38f8-42ea-b6be-00cc1da8c766 req-4ca5e16e-c313-495d-b2c2-46426958694f service nova] Lock "e845f60b-004f-4a5c-813e-0c183e99eba6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 819.182218] env[63088]: DEBUG oslo_concurrency.lockutils [req-2e64b50e-38f8-42ea-b6be-00cc1da8c766 req-4ca5e16e-c313-495d-b2c2-46426958694f service nova] Lock "e845f60b-004f-4a5c-813e-0c183e99eba6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 819.182366] env[63088]: DEBUG nova.compute.manager [req-2e64b50e-38f8-42ea-b6be-00cc1da8c766 req-4ca5e16e-c313-495d-b2c2-46426958694f service nova] [instance: e845f60b-004f-4a5c-813e-0c183e99eba6] No waiting events found dispatching network-vif-plugged-8ca1e82d-550d-4faa-8b7e-93e0b812c234 {{(pid=63088) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 819.182539] env[63088]: WARNING nova.compute.manager [req-2e64b50e-38f8-42ea-b6be-00cc1da8c766 req-4ca5e16e-c313-495d-b2c2-46426958694f service nova] [instance: e845f60b-004f-4a5c-813e-0c183e99eba6] Received unexpected event network-vif-plugged-8ca1e82d-550d-4faa-8b7e-93e0b812c234 for instance with vm_state building and task_state spawning. [ 819.182711] env[63088]: DEBUG nova.compute.manager [req-2e64b50e-38f8-42ea-b6be-00cc1da8c766 req-4ca5e16e-c313-495d-b2c2-46426958694f service nova] [instance: e845f60b-004f-4a5c-813e-0c183e99eba6] Received event network-changed-8ca1e82d-550d-4faa-8b7e-93e0b812c234 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 819.182932] env[63088]: DEBUG nova.compute.manager [req-2e64b50e-38f8-42ea-b6be-00cc1da8c766 req-4ca5e16e-c313-495d-b2c2-46426958694f service nova] [instance: e845f60b-004f-4a5c-813e-0c183e99eba6] Refreshing instance network info cache due to event network-changed-8ca1e82d-550d-4faa-8b7e-93e0b812c234. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 819.183057] env[63088]: DEBUG oslo_concurrency.lockutils [req-2e64b50e-38f8-42ea-b6be-00cc1da8c766 req-4ca5e16e-c313-495d-b2c2-46426958694f service nova] Acquiring lock "refresh_cache-e845f60b-004f-4a5c-813e-0c183e99eba6" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 819.270361] env[63088]: DEBUG oslo_concurrency.lockutils [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Releasing lock "refresh_cache-e845f60b-004f-4a5c-813e-0c183e99eba6" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 819.270699] env[63088]: DEBUG nova.compute.manager [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: e845f60b-004f-4a5c-813e-0c183e99eba6] Instance network_info: |[{"id": "8ca1e82d-550d-4faa-8b7e-93e0b812c234", "address": "fa:16:3e:11:39:ef", "network": {"id": "307b966a-d9e1-40fd-9313-1ad94c734308", "bridge": "br-int", "label": "tempest-ServersTestJSON-1907249197-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "03e20d606e654362acbe2b36fe499ae3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae18b41f-e73c-44f1-83dd-467c080944f4", "external-id": "nsx-vlan-transportzone-653", "segmentation_id": 653, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8ca1e82d-55", "ovs_interfaceid": "8ca1e82d-550d-4faa-8b7e-93e0b812c234", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 819.271012] env[63088]: DEBUG oslo_concurrency.lockutils [req-2e64b50e-38f8-42ea-b6be-00cc1da8c766 req-4ca5e16e-c313-495d-b2c2-46426958694f service nova] Acquired lock "refresh_cache-e845f60b-004f-4a5c-813e-0c183e99eba6" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 819.271202] env[63088]: DEBUG nova.network.neutron [req-2e64b50e-38f8-42ea-b6be-00cc1da8c766 req-4ca5e16e-c313-495d-b2c2-46426958694f service nova] [instance: e845f60b-004f-4a5c-813e-0c183e99eba6] Refreshing network info cache for port 8ca1e82d-550d-4faa-8b7e-93e0b812c234 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 819.272565] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: e845f60b-004f-4a5c-813e-0c183e99eba6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:11:39:ef', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ae18b41f-e73c-44f1-83dd-467c080944f4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8ca1e82d-550d-4faa-8b7e-93e0b812c234', 'vif_model': 'vmxnet3'}] {{(pid=63088) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 819.281886] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Creating folder: Project (03e20d606e654362acbe2b36fe499ae3). Parent ref: group-v275816. {{(pid=63088) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 819.282963] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4080048b-0c1d-4c8b-a191-530c03bd2dd4 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.295940] env[63088]: INFO nova.virt.vmwareapi.vm_util [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Created folder: Project (03e20d606e654362acbe2b36fe499ae3) in parent group-v275816. [ 819.296172] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Creating folder: Instances. Parent ref: group-v275845. {{(pid=63088) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 819.296425] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3ad20a2e-46f2-456e-8d68-e248bc95916b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.307120] env[63088]: INFO nova.virt.vmwareapi.vm_util [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Created folder: Instances in parent group-v275845. [ 819.307409] env[63088]: DEBUG oslo.service.loopingcall [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 819.307749] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e845f60b-004f-4a5c-813e-0c183e99eba6] Creating VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 819.307951] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-15c59a2d-7462-4ae2-a215-35b7f1d0bb57 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.324683] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.911s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 819.325207] env[63088]: DEBUG nova.compute.manager [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 819.327803] env[63088]: DEBUG oslo_concurrency.lockutils [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.678s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 819.329278] env[63088]: INFO nova.compute.claims [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] [instance: b159649b-ebd5-4c7d-9074-dc0e6395a947] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 819.337037] env[63088]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 819.337037] env[63088]: value = "task-1284865" [ 819.337037] env[63088]: _type = "Task" [ 819.337037] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.345371] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1284865, 'name': CreateVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.411420] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-69612416-d1eb-4a97-8ff2-4c55509ccebf tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: db032da3-d334-486f-a428-af1c8a3c360a] VM already powered off {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 819.411717] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-69612416-d1eb-4a97-8ff2-4c55509ccebf tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: db032da3-d334-486f-a428-af1c8a3c360a] Processing image 1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 819.411958] env[63088]: DEBUG oslo_concurrency.lockutils [None req-69612416-d1eb-4a97-8ff2-4c55509ccebf tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 819.412160] env[63088]: DEBUG oslo_concurrency.lockutils [None req-69612416-d1eb-4a97-8ff2-4c55509ccebf tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 819.412340] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-69612416-d1eb-4a97-8ff2-4c55509ccebf tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 819.412640] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fe49678f-2a38-4ae9-9568-0ad6fa52c191 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.423068] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-69612416-d1eb-4a97-8ff2-4c55509ccebf tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 819.423269] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-69612416-d1eb-4a97-8ff2-4c55509ccebf tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 819.423919] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ed39bfc2-3ac8-4b9d-9220-d20cf66d9e4e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.430295] env[63088]: DEBUG oslo_vmware.api [None req-69612416-d1eb-4a97-8ff2-4c55509ccebf tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Waiting for the task: (returnval){ [ 819.430295] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]520cf26a-e991-2d41-be2f-b56065ab37e0" [ 819.430295] env[63088]: _type = "Task" [ 819.430295] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.439853] env[63088]: DEBUG oslo_vmware.api [None req-69612416-d1eb-4a97-8ff2-4c55509ccebf tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]520cf26a-e991-2d41-be2f-b56065ab37e0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.444230] env[63088]: DEBUG nova.compute.manager [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] [instance: 4f6e2d99-7300-42e6-8abf-362ddd4e8c6d] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 819.470486] env[63088]: DEBUG nova.virt.hardware [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 819.470931] env[63088]: DEBUG nova.virt.hardware [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 819.471166] env[63088]: DEBUG nova.virt.hardware [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 819.471373] env[63088]: DEBUG nova.virt.hardware [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 819.471524] env[63088]: DEBUG nova.virt.hardware [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 819.471675] env[63088]: DEBUG nova.virt.hardware [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 819.471896] env[63088]: DEBUG nova.virt.hardware [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 819.472074] env[63088]: DEBUG nova.virt.hardware [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 819.472669] env[63088]: DEBUG nova.virt.hardware [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 819.472898] env[63088]: DEBUG nova.virt.hardware [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 819.473058] env[63088]: DEBUG nova.virt.hardware [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 819.473979] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfc69622-ce84-4a9d-8b14-564f8589f7b5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.483474] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a9745cb-07e3-416f-9a97-562a2d7de47b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.518359] env[63088]: DEBUG oslo_concurrency.lockutils [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 819.562610] env[63088]: DEBUG oslo_vmware.api [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1284860, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.657997] env[63088]: DEBUG oslo_vmware.api [None req-da2e38c8-9951-4dcc-950e-b6d060a5ab7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1284861, 'name': PowerOffVM_Task, 'duration_secs': 0.189955} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.658295] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-da2e38c8-9951-4dcc-950e-b6d060a5ab7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 11ec987f-9908-4366-81ae-cbfe4de67b2d] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 819.658464] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-da2e38c8-9951-4dcc-950e-b6d060a5ab7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 11ec987f-9908-4366-81ae-cbfe4de67b2d] Unregistering the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 819.658714] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-237ce61f-4c54-452f-a72c-88efbef94eb1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.683258] env[63088]: DEBUG oslo_vmware.api [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1284862, 'name': PowerOnVM_Task} progress is 19%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.716299] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-da2e38c8-9951-4dcc-950e-b6d060a5ab7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 11ec987f-9908-4366-81ae-cbfe4de67b2d] Unregistered the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 819.716512] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-da2e38c8-9951-4dcc-950e-b6d060a5ab7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 11ec987f-9908-4366-81ae-cbfe4de67b2d] Deleting contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 819.716669] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-da2e38c8-9951-4dcc-950e-b6d060a5ab7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Deleting the datastore file [datastore1] 11ec987f-9908-4366-81ae-cbfe4de67b2d {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 819.716940] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fd664b3c-b734-48af-adae-775699f98e1e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.723733] env[63088]: DEBUG oslo_vmware.api [None req-da2e38c8-9951-4dcc-950e-b6d060a5ab7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 819.723733] env[63088]: value = "task-1284867" [ 819.723733] env[63088]: _type = "Task" [ 819.723733] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.733263] env[63088]: DEBUG oslo_vmware.api [None req-da2e38c8-9951-4dcc-950e-b6d060a5ab7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1284867, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.834273] env[63088]: DEBUG nova.compute.utils [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 819.837681] env[63088]: DEBUG nova.compute.manager [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 819.837853] env[63088]: DEBUG nova.network.neutron [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 819.852999] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1284865, 'name': CreateVM_Task, 'duration_secs': 0.46476} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.853261] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e845f60b-004f-4a5c-813e-0c183e99eba6] Created VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 819.854398] env[63088]: DEBUG oslo_concurrency.lockutils [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 819.854666] env[63088]: DEBUG oslo_concurrency.lockutils [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 819.855140] env[63088]: DEBUG oslo_concurrency.lockutils [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 819.855792] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b76299a0-4f1d-4b1e-9387-59d13ba76459 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.862054] env[63088]: DEBUG oslo_vmware.api [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Waiting for the task: (returnval){ [ 819.862054] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52e4bc4d-91ef-56a0-28c6-f68046530046" [ 819.862054] env[63088]: _type = "Task" [ 819.862054] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.873820] env[63088]: DEBUG oslo_vmware.api [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52e4bc4d-91ef-56a0-28c6-f68046530046, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.920500] env[63088]: DEBUG nova.network.neutron [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] [instance: 4f6e2d99-7300-42e6-8abf-362ddd4e8c6d] Successfully updated port: 07c78e97-0b3c-45a7-859f-bbf74ce3c5f6 {{(pid=63088) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 819.941569] env[63088]: DEBUG oslo_vmware.api [None req-69612416-d1eb-4a97-8ff2-4c55509ccebf tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]520cf26a-e991-2d41-be2f-b56065ab37e0, 'name': SearchDatastore_Task, 'duration_secs': 0.014377} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.942996] env[63088]: DEBUG nova.policy [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3485039c4d464c0582ffcd6d4eaac40b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '937eba1a40c84534b279d5798e234625', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 819.945204] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-37960d02-7af0-421d-91f6-d214f58f00bd {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.951407] env[63088]: DEBUG oslo_vmware.api [None req-69612416-d1eb-4a97-8ff2-4c55509ccebf tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Waiting for the task: (returnval){ [ 819.951407] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]5269164c-2c73-22af-5fc1-9c60da580b22" [ 819.951407] env[63088]: _type = "Task" [ 819.951407] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.959650] env[63088]: DEBUG oslo_vmware.api [None req-69612416-d1eb-4a97-8ff2-4c55509ccebf tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]5269164c-2c73-22af-5fc1-9c60da580b22, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.064506] env[63088]: DEBUG oslo_vmware.api [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1284860, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.573397} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.064831] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore2] 4847f15e-98d4-401c-91ff-067e84a85727/4847f15e-98d4-401c-91ff-067e84a85727.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 820.065082] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] Extending root virtual disk to 1048576 {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 820.065416] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1bb2d165-5a1e-4f9f-a3bf-83f0be3c3fc5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.072484] env[63088]: DEBUG oslo_vmware.api [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Waiting for the task: (returnval){ [ 820.072484] env[63088]: value = "task-1284868" [ 820.072484] env[63088]: _type = "Task" [ 820.072484] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.081171] env[63088]: DEBUG oslo_vmware.api [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1284868, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.166232] env[63088]: DEBUG nova.network.neutron [req-2e64b50e-38f8-42ea-b6be-00cc1da8c766 req-4ca5e16e-c313-495d-b2c2-46426958694f service nova] [instance: e845f60b-004f-4a5c-813e-0c183e99eba6] Updated VIF entry in instance network info cache for port 8ca1e82d-550d-4faa-8b7e-93e0b812c234. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 820.166629] env[63088]: DEBUG nova.network.neutron [req-2e64b50e-38f8-42ea-b6be-00cc1da8c766 req-4ca5e16e-c313-495d-b2c2-46426958694f service nova] [instance: e845f60b-004f-4a5c-813e-0c183e99eba6] Updating instance_info_cache with network_info: [{"id": "8ca1e82d-550d-4faa-8b7e-93e0b812c234", "address": "fa:16:3e:11:39:ef", "network": {"id": "307b966a-d9e1-40fd-9313-1ad94c734308", "bridge": "br-int", "label": "tempest-ServersTestJSON-1907249197-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "03e20d606e654362acbe2b36fe499ae3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae18b41f-e73c-44f1-83dd-467c080944f4", "external-id": "nsx-vlan-transportzone-653", "segmentation_id": 653, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8ca1e82d-55", "ovs_interfaceid": "8ca1e82d-550d-4faa-8b7e-93e0b812c234", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 820.183288] env[63088]: DEBUG oslo_vmware.api [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1284862, 'name': PowerOnVM_Task, 'duration_secs': 0.990686} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.183538] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 820.183737] env[63088]: INFO nova.compute.manager [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Took 8.47 seconds to spawn the instance on the hypervisor. [ 820.183918] env[63088]: DEBUG nova.compute.manager [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 820.184662] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec94f552-186f-4a68-ae24-7786b0044bf7 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.233815] env[63088]: DEBUG oslo_vmware.api [None req-da2e38c8-9951-4dcc-950e-b6d060a5ab7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1284867, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.147152} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.234156] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-da2e38c8-9951-4dcc-950e-b6d060a5ab7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 820.234363] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-da2e38c8-9951-4dcc-950e-b6d060a5ab7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 11ec987f-9908-4366-81ae-cbfe4de67b2d] Deleted contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 820.234566] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-da2e38c8-9951-4dcc-950e-b6d060a5ab7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 11ec987f-9908-4366-81ae-cbfe4de67b2d] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 820.234739] env[63088]: INFO nova.compute.manager [None req-da2e38c8-9951-4dcc-950e-b6d060a5ab7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 11ec987f-9908-4366-81ae-cbfe4de67b2d] Took 1.10 seconds to destroy the instance on the hypervisor. [ 820.234974] env[63088]: DEBUG oslo.service.loopingcall [None req-da2e38c8-9951-4dcc-950e-b6d060a5ab7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 820.235183] env[63088]: DEBUG nova.compute.manager [-] [instance: 11ec987f-9908-4366-81ae-cbfe4de67b2d] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 820.235276] env[63088]: DEBUG nova.network.neutron [-] [instance: 11ec987f-9908-4366-81ae-cbfe4de67b2d] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 820.338820] env[63088]: DEBUG nova.compute.manager [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 820.381965] env[63088]: DEBUG oslo_vmware.api [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52e4bc4d-91ef-56a0-28c6-f68046530046, 'name': SearchDatastore_Task, 'duration_secs': 0.010833} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.382276] env[63088]: DEBUG oslo_concurrency.lockutils [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 820.382501] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: e845f60b-004f-4a5c-813e-0c183e99eba6] Processing image 1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 820.382728] env[63088]: DEBUG oslo_concurrency.lockutils [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 820.382871] env[63088]: DEBUG oslo_concurrency.lockutils [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 820.383053] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 820.383312] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7e3d9786-7e6e-4f24-a9b0-b3aaea623975 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.393707] env[63088]: DEBUG nova.network.neutron [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Successfully created port: 0c182796-7fb6-40ab-ba27-a43f0700dee1 {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 820.398272] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 820.402896] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 820.402896] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8cc93e21-5d9d-4ef3-984b-dcb0adc2c3d9 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.408121] env[63088]: DEBUG oslo_vmware.api [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Waiting for the task: (returnval){ [ 820.408121] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52186e88-42bb-1545-9224-0506402d4c1a" [ 820.408121] env[63088]: _type = "Task" [ 820.408121] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.415851] env[63088]: DEBUG oslo_vmware.api [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52186e88-42bb-1545-9224-0506402d4c1a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.419695] env[63088]: DEBUG oslo_concurrency.lockutils [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Acquiring lock "refresh_cache-4f6e2d99-7300-42e6-8abf-362ddd4e8c6d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 820.420173] env[63088]: DEBUG oslo_concurrency.lockutils [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Acquired lock "refresh_cache-4f6e2d99-7300-42e6-8abf-362ddd4e8c6d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 820.420173] env[63088]: DEBUG nova.network.neutron [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] [instance: 4f6e2d99-7300-42e6-8abf-362ddd4e8c6d] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 820.463355] env[63088]: DEBUG oslo_vmware.api [None req-69612416-d1eb-4a97-8ff2-4c55509ccebf tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]5269164c-2c73-22af-5fc1-9c60da580b22, 'name': SearchDatastore_Task, 'duration_secs': 0.00917} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.463621] env[63088]: DEBUG oslo_concurrency.lockutils [None req-69612416-d1eb-4a97-8ff2-4c55509ccebf tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 820.463842] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-69612416-d1eb-4a97-8ff2-4c55509ccebf tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] db032da3-d334-486f-a428-af1c8a3c360a/1e8c5d18-0a03-4e18-afe1-de5a6e255953-rescue.vmdk. {{(pid=63088) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 820.464394] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bc5389f1-37c7-422e-a19c-25988f9eee9a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.474172] env[63088]: DEBUG oslo_vmware.api [None req-69612416-d1eb-4a97-8ff2-4c55509ccebf tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Waiting for the task: (returnval){ [ 820.474172] env[63088]: value = "task-1284869" [ 820.474172] env[63088]: _type = "Task" [ 820.474172] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.482927] env[63088]: DEBUG oslo_vmware.api [None req-69612416-d1eb-4a97-8ff2-4c55509ccebf tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': task-1284869, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.585145] env[63088]: DEBUG oslo_vmware.api [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1284868, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064373} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.585441] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] Extended root virtual disk {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 820.586274] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d668ef8a-17ed-4729-a862-5178ed637735 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.610516] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] Reconfiguring VM instance instance-00000037 to attach disk [datastore2] 4847f15e-98d4-401c-91ff-067e84a85727/4847f15e-98d4-401c-91ff-067e84a85727.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 820.612167] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6cb8ba06-267b-49ec-b67f-c0b378aca6f6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.628090] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc893957-3d7f-498b-9910-8feac1f2e1c8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.639024] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c67eb16-ddc8-4c5d-ad6f-fa763ae0e8e2 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.640584] env[63088]: DEBUG oslo_vmware.api [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Waiting for the task: (returnval){ [ 820.640584] env[63088]: value = "task-1284870" [ 820.640584] env[63088]: _type = "Task" [ 820.640584] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.670917] env[63088]: DEBUG oslo_concurrency.lockutils [req-2e64b50e-38f8-42ea-b6be-00cc1da8c766 req-4ca5e16e-c313-495d-b2c2-46426958694f service nova] Releasing lock "refresh_cache-e845f60b-004f-4a5c-813e-0c183e99eba6" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 820.676510] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77341e9b-12a7-4d47-889f-c4d9e7f64c75 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.679280] env[63088]: DEBUG oslo_vmware.api [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1284870, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.684718] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18ac3017-b99f-43f4-81ba-818ac7351da3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.710684] env[63088]: DEBUG nova.compute.provider_tree [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 820.714270] env[63088]: INFO nova.compute.manager [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Took 36.74 seconds to build instance. [ 820.733067] env[63088]: DEBUG nova.compute.manager [req-d03bc2b8-851a-4934-88dc-4914f8eb5382 req-00a13999-b3bd-4e5e-9226-895c9d342cae service nova] [instance: 11ec987f-9908-4366-81ae-cbfe4de67b2d] Received event network-vif-deleted-90ea719e-0a3d-4bae-a472-a58cbb3c1541 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 820.733272] env[63088]: INFO nova.compute.manager [req-d03bc2b8-851a-4934-88dc-4914f8eb5382 req-00a13999-b3bd-4e5e-9226-895c9d342cae service nova] [instance: 11ec987f-9908-4366-81ae-cbfe4de67b2d] Neutron deleted interface 90ea719e-0a3d-4bae-a472-a58cbb3c1541; detaching it from the instance and deleting it from the info cache [ 820.733444] env[63088]: DEBUG nova.network.neutron [req-d03bc2b8-851a-4934-88dc-4914f8eb5382 req-00a13999-b3bd-4e5e-9226-895c9d342cae service nova] [instance: 11ec987f-9908-4366-81ae-cbfe4de67b2d] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 820.919691] env[63088]: DEBUG oslo_vmware.api [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52186e88-42bb-1545-9224-0506402d4c1a, 'name': SearchDatastore_Task, 'duration_secs': 0.007911} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.920638] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9c0118da-10cd-4ca7-a022-b055c3877c21 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.928505] env[63088]: DEBUG oslo_vmware.api [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Waiting for the task: (returnval){ [ 820.928505] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52e7aa9d-dd0b-a344-4dc7-e8ef7266e501" [ 820.928505] env[63088]: _type = "Task" [ 820.928505] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.938041] env[63088]: DEBUG oslo_vmware.api [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52e7aa9d-dd0b-a344-4dc7-e8ef7266e501, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.973229] env[63088]: DEBUG nova.network.neutron [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] [instance: 4f6e2d99-7300-42e6-8abf-362ddd4e8c6d] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 820.983801] env[63088]: DEBUG oslo_vmware.api [None req-69612416-d1eb-4a97-8ff2-4c55509ccebf tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': task-1284869, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.504049} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.984093] env[63088]: INFO nova.virt.vmwareapi.ds_util [None req-69612416-d1eb-4a97-8ff2-4c55509ccebf tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] db032da3-d334-486f-a428-af1c8a3c360a/1e8c5d18-0a03-4e18-afe1-de5a6e255953-rescue.vmdk. [ 820.984864] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10be714b-4cf0-4927-85e8-da0af719d3fe {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.011030] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-69612416-d1eb-4a97-8ff2-4c55509ccebf tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: db032da3-d334-486f-a428-af1c8a3c360a] Reconfiguring VM instance instance-00000034 to attach disk [datastore1] db032da3-d334-486f-a428-af1c8a3c360a/1e8c5d18-0a03-4e18-afe1-de5a6e255953-rescue.vmdk or device None with type thin {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 821.011288] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-039363f2-b45f-4e66-a47e-8acf50b411b3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.028252] env[63088]: DEBUG oslo_vmware.api [None req-69612416-d1eb-4a97-8ff2-4c55509ccebf tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Waiting for the task: (returnval){ [ 821.028252] env[63088]: value = "task-1284871" [ 821.028252] env[63088]: _type = "Task" [ 821.028252] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.038203] env[63088]: DEBUG oslo_vmware.api [None req-69612416-d1eb-4a97-8ff2-4c55509ccebf tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': task-1284871, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.152399] env[63088]: DEBUG oslo_vmware.api [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1284870, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.161692] env[63088]: DEBUG nova.network.neutron [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] [instance: 4f6e2d99-7300-42e6-8abf-362ddd4e8c6d] Updating instance_info_cache with network_info: [{"id": "07c78e97-0b3c-45a7-859f-bbf74ce3c5f6", "address": "fa:16:3e:8f:bb:e9", "network": {"id": "5990b541-319a-48bf-a47f-70452bdab9ce", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-1642910243-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "960b3ebce4b747dfa84c1659118c78f0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bec1528b-3e87-477b-8ab2-02696ad47e66", "external-id": "nsx-vlan-transportzone-180", "segmentation_id": 180, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap07c78e97-0b", "ovs_interfaceid": "07c78e97-0b3c-45a7-859f-bbf74ce3c5f6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 821.180486] env[63088]: DEBUG nova.network.neutron [-] [instance: 11ec987f-9908-4366-81ae-cbfe4de67b2d] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 821.215093] env[63088]: DEBUG nova.scheduler.client.report [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 821.218702] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bd64f004-6ec5-4165-9652-c6aeea865eb4 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Lock "4971b24c-6710-4f50-9846-727dad264b1f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 79.542s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.238674] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-617d9c41-0a4b-480a-a472-1012648a8fe3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.249601] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1855e5e4-f2a7-47a4-b8ab-99935eb8c01b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.274427] env[63088]: DEBUG nova.compute.manager [req-d03bc2b8-851a-4934-88dc-4914f8eb5382 req-00a13999-b3bd-4e5e-9226-895c9d342cae service nova] [instance: 11ec987f-9908-4366-81ae-cbfe4de67b2d] Detach interface failed, port_id=90ea719e-0a3d-4bae-a472-a58cbb3c1541, reason: Instance 11ec987f-9908-4366-81ae-cbfe4de67b2d could not be found. {{(pid=63088) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 821.354275] env[63088]: DEBUG nova.compute.manager [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 821.364714] env[63088]: DEBUG nova.compute.manager [req-4bec4a38-204a-48d3-b7b8-9d362d3c27b2 req-ad4dbd71-0f93-4360-a794-9052ff60ff7e service nova] [instance: 4f6e2d99-7300-42e6-8abf-362ddd4e8c6d] Received event network-vif-plugged-07c78e97-0b3c-45a7-859f-bbf74ce3c5f6 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 821.364915] env[63088]: DEBUG oslo_concurrency.lockutils [req-4bec4a38-204a-48d3-b7b8-9d362d3c27b2 req-ad4dbd71-0f93-4360-a794-9052ff60ff7e service nova] Acquiring lock "4f6e2d99-7300-42e6-8abf-362ddd4e8c6d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 821.366457] env[63088]: DEBUG oslo_concurrency.lockutils [req-4bec4a38-204a-48d3-b7b8-9d362d3c27b2 req-ad4dbd71-0f93-4360-a794-9052ff60ff7e service nova] Lock "4f6e2d99-7300-42e6-8abf-362ddd4e8c6d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.366457] env[63088]: DEBUG oslo_concurrency.lockutils [req-4bec4a38-204a-48d3-b7b8-9d362d3c27b2 req-ad4dbd71-0f93-4360-a794-9052ff60ff7e service nova] Lock "4f6e2d99-7300-42e6-8abf-362ddd4e8c6d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.366457] env[63088]: DEBUG nova.compute.manager [req-4bec4a38-204a-48d3-b7b8-9d362d3c27b2 req-ad4dbd71-0f93-4360-a794-9052ff60ff7e service nova] [instance: 4f6e2d99-7300-42e6-8abf-362ddd4e8c6d] No waiting events found dispatching network-vif-plugged-07c78e97-0b3c-45a7-859f-bbf74ce3c5f6 {{(pid=63088) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 821.366457] env[63088]: WARNING nova.compute.manager [req-4bec4a38-204a-48d3-b7b8-9d362d3c27b2 req-ad4dbd71-0f93-4360-a794-9052ff60ff7e service nova] [instance: 4f6e2d99-7300-42e6-8abf-362ddd4e8c6d] Received unexpected event network-vif-plugged-07c78e97-0b3c-45a7-859f-bbf74ce3c5f6 for instance with vm_state building and task_state spawning. [ 821.366457] env[63088]: DEBUG nova.compute.manager [req-4bec4a38-204a-48d3-b7b8-9d362d3c27b2 req-ad4dbd71-0f93-4360-a794-9052ff60ff7e service nova] [instance: 4f6e2d99-7300-42e6-8abf-362ddd4e8c6d] Received event network-changed-07c78e97-0b3c-45a7-859f-bbf74ce3c5f6 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 821.366607] env[63088]: DEBUG nova.compute.manager [req-4bec4a38-204a-48d3-b7b8-9d362d3c27b2 req-ad4dbd71-0f93-4360-a794-9052ff60ff7e service nova] [instance: 4f6e2d99-7300-42e6-8abf-362ddd4e8c6d] Refreshing instance network info cache due to event network-changed-07c78e97-0b3c-45a7-859f-bbf74ce3c5f6. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 821.366607] env[63088]: DEBUG oslo_concurrency.lockutils [req-4bec4a38-204a-48d3-b7b8-9d362d3c27b2 req-ad4dbd71-0f93-4360-a794-9052ff60ff7e service nova] Acquiring lock "refresh_cache-4f6e2d99-7300-42e6-8abf-362ddd4e8c6d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 821.381657] env[63088]: DEBUG nova.virt.hardware [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 821.381853] env[63088]: DEBUG nova.virt.hardware [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 821.382026] env[63088]: DEBUG nova.virt.hardware [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 821.382194] env[63088]: DEBUG nova.virt.hardware [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 821.382338] env[63088]: DEBUG nova.virt.hardware [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 821.382474] env[63088]: DEBUG nova.virt.hardware [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 821.382859] env[63088]: DEBUG nova.virt.hardware [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 821.382859] env[63088]: DEBUG nova.virt.hardware [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 821.382972] env[63088]: DEBUG nova.virt.hardware [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 821.383135] env[63088]: DEBUG nova.virt.hardware [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 821.383307] env[63088]: DEBUG nova.virt.hardware [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 821.384402] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33c851ba-2b8d-43cf-92d0-48f4768457b7 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.392342] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3d6d836-1e0f-473c-8937-6e5c06365b84 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.437746] env[63088]: DEBUG oslo_vmware.api [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52e7aa9d-dd0b-a344-4dc7-e8ef7266e501, 'name': SearchDatastore_Task, 'duration_secs': 0.018661} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.438365] env[63088]: DEBUG oslo_concurrency.lockutils [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 821.438627] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore2] e845f60b-004f-4a5c-813e-0c183e99eba6/e845f60b-004f-4a5c-813e-0c183e99eba6.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 821.438882] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fce95c8d-17c3-481c-a974-f961ebbd3b95 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.445698] env[63088]: DEBUG oslo_vmware.api [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Waiting for the task: (returnval){ [ 821.445698] env[63088]: value = "task-1284872" [ 821.445698] env[63088]: _type = "Task" [ 821.445698] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.453192] env[63088]: DEBUG oslo_vmware.api [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1284872, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.537723] env[63088]: DEBUG oslo_vmware.api [None req-69612416-d1eb-4a97-8ff2-4c55509ccebf tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': task-1284871, 'name': ReconfigVM_Task, 'duration_secs': 0.274975} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.538058] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-69612416-d1eb-4a97-8ff2-4c55509ccebf tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: db032da3-d334-486f-a428-af1c8a3c360a] Reconfigured VM instance instance-00000034 to attach disk [datastore1] db032da3-d334-486f-a428-af1c8a3c360a/1e8c5d18-0a03-4e18-afe1-de5a6e255953-rescue.vmdk or device None with type thin {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 821.538905] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3dd2795-93a9-42a6-987c-8d340809a7da {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.563600] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2c172b5c-6e5b-40f9-866a-40d74a261b66 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.578889] env[63088]: DEBUG oslo_vmware.api [None req-69612416-d1eb-4a97-8ff2-4c55509ccebf tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Waiting for the task: (returnval){ [ 821.578889] env[63088]: value = "task-1284873" [ 821.578889] env[63088]: _type = "Task" [ 821.578889] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.589425] env[63088]: DEBUG oslo_vmware.api [None req-69612416-d1eb-4a97-8ff2-4c55509ccebf tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': task-1284873, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.651837] env[63088]: DEBUG oslo_vmware.api [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1284870, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.666625] env[63088]: DEBUG oslo_concurrency.lockutils [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Releasing lock "refresh_cache-4f6e2d99-7300-42e6-8abf-362ddd4e8c6d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 821.666979] env[63088]: DEBUG nova.compute.manager [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] [instance: 4f6e2d99-7300-42e6-8abf-362ddd4e8c6d] Instance network_info: |[{"id": "07c78e97-0b3c-45a7-859f-bbf74ce3c5f6", "address": "fa:16:3e:8f:bb:e9", "network": {"id": "5990b541-319a-48bf-a47f-70452bdab9ce", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-1642910243-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "960b3ebce4b747dfa84c1659118c78f0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bec1528b-3e87-477b-8ab2-02696ad47e66", "external-id": "nsx-vlan-transportzone-180", "segmentation_id": 180, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap07c78e97-0b", "ovs_interfaceid": "07c78e97-0b3c-45a7-859f-bbf74ce3c5f6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 821.667418] env[63088]: DEBUG oslo_concurrency.lockutils [req-4bec4a38-204a-48d3-b7b8-9d362d3c27b2 req-ad4dbd71-0f93-4360-a794-9052ff60ff7e service nova] Acquired lock "refresh_cache-4f6e2d99-7300-42e6-8abf-362ddd4e8c6d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 821.667690] env[63088]: DEBUG nova.network.neutron [req-4bec4a38-204a-48d3-b7b8-9d362d3c27b2 req-ad4dbd71-0f93-4360-a794-9052ff60ff7e service nova] [instance: 4f6e2d99-7300-42e6-8abf-362ddd4e8c6d] Refreshing network info cache for port 07c78e97-0b3c-45a7-859f-bbf74ce3c5f6 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 821.669054] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] [instance: 4f6e2d99-7300-42e6-8abf-362ddd4e8c6d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8f:bb:e9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bec1528b-3e87-477b-8ab2-02696ad47e66', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '07c78e97-0b3c-45a7-859f-bbf74ce3c5f6', 'vif_model': 'vmxnet3'}] {{(pid=63088) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 821.676771] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Creating folder: Project (960b3ebce4b747dfa84c1659118c78f0). Parent ref: group-v275816. {{(pid=63088) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 821.677399] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1da6f20b-6df7-4c8b-a279-1a4fa881e8f6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.683487] env[63088]: INFO nova.compute.manager [-] [instance: 11ec987f-9908-4366-81ae-cbfe4de67b2d] Took 1.45 seconds to deallocate network for instance. [ 821.690076] env[63088]: INFO nova.virt.vmwareapi.vm_util [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Created folder: Project (960b3ebce4b747dfa84c1659118c78f0) in parent group-v275816. [ 821.690302] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Creating folder: Instances. Parent ref: group-v275848. {{(pid=63088) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 821.690722] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-15588b39-5a65-448e-b2dd-16fcd4724c65 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.700427] env[63088]: INFO nova.virt.vmwareapi.vm_util [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Created folder: Instances in parent group-v275848. [ 821.700774] env[63088]: DEBUG oslo.service.loopingcall [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 821.700903] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4f6e2d99-7300-42e6-8abf-362ddd4e8c6d] Creating VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 821.701129] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-df2c0974-e51e-495b-b5e7-1898c8d6aa3e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.720431] env[63088]: DEBUG oslo_concurrency.lockutils [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.393s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.721598] env[63088]: DEBUG nova.compute.manager [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] [instance: b159649b-ebd5-4c7d-9074-dc0e6395a947] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 821.725335] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b5c531a2-a229-4998-9a9b-9ebb1e7afff2 tempest-ServerGroupTestJSON-1534054659 tempest-ServerGroupTestJSON-1534054659-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 25.985s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.729289] env[63088]: DEBUG nova.compute.manager [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 00925f94-dbf9-453c-a124-d8434679aedf] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 821.730698] env[63088]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 821.730698] env[63088]: value = "task-1284876" [ 821.730698] env[63088]: _type = "Task" [ 821.730698] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.742553] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1284876, 'name': CreateVM_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.956145] env[63088]: DEBUG oslo_vmware.api [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1284872, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.460112} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.956496] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore2] e845f60b-004f-4a5c-813e-0c183e99eba6/e845f60b-004f-4a5c-813e-0c183e99eba6.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 821.957278] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: e845f60b-004f-4a5c-813e-0c183e99eba6] Extending root virtual disk to 1048576 {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 821.957278] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-57182c00-2d91-4ed0-80bb-aa3632daf03a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.963596] env[63088]: DEBUG oslo_vmware.api [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Waiting for the task: (returnval){ [ 821.963596] env[63088]: value = "task-1284877" [ 821.963596] env[63088]: _type = "Task" [ 821.963596] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.972102] env[63088]: DEBUG oslo_vmware.api [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1284877, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.094886] env[63088]: DEBUG oslo_vmware.api [None req-69612416-d1eb-4a97-8ff2-4c55509ccebf tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': task-1284873, 'name': ReconfigVM_Task, 'duration_secs': 0.166543} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.095327] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-69612416-d1eb-4a97-8ff2-4c55509ccebf tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: db032da3-d334-486f-a428-af1c8a3c360a] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 822.095676] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-23975d94-be2b-4bda-9454-c65b434fc20e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.102876] env[63088]: DEBUG oslo_vmware.api [None req-69612416-d1eb-4a97-8ff2-4c55509ccebf tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Waiting for the task: (returnval){ [ 822.102876] env[63088]: value = "task-1284878" [ 822.102876] env[63088]: _type = "Task" [ 822.102876] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.120213] env[63088]: DEBUG oslo_vmware.api [None req-69612416-d1eb-4a97-8ff2-4c55509ccebf tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': task-1284878, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.155253] env[63088]: DEBUG oslo_vmware.api [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1284870, 'name': ReconfigVM_Task, 'duration_secs': 1.493014} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.155648] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] Reconfigured VM instance instance-00000037 to attach disk [datastore2] 4847f15e-98d4-401c-91ff-067e84a85727/4847f15e-98d4-401c-91ff-067e84a85727.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 822.156527] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4d237129-f93a-4e3a-9055-3417b151a153 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.164253] env[63088]: DEBUG oslo_vmware.api [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Waiting for the task: (returnval){ [ 822.164253] env[63088]: value = "task-1284879" [ 822.164253] env[63088]: _type = "Task" [ 822.164253] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.174682] env[63088]: DEBUG oslo_vmware.api [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1284879, 'name': Rename_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.193710] env[63088]: DEBUG oslo_concurrency.lockutils [None req-da2e38c8-9951-4dcc-950e-b6d060a5ab7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 822.203587] env[63088]: DEBUG nova.network.neutron [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Successfully updated port: 0c182796-7fb6-40ab-ba27-a43f0700dee1 {{(pid=63088) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 822.233317] env[63088]: DEBUG nova.compute.utils [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 822.238946] env[63088]: DEBUG nova.compute.manager [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] [instance: b159649b-ebd5-4c7d-9074-dc0e6395a947] Not allocating networking since 'none' was specified. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 822.252616] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1284876, 'name': CreateVM_Task, 'duration_secs': 0.415472} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.252761] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4f6e2d99-7300-42e6-8abf-362ddd4e8c6d] Created VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 822.253937] env[63088]: DEBUG oslo_concurrency.lockutils [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 822.254313] env[63088]: DEBUG oslo_concurrency.lockutils [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 822.254432] env[63088]: DEBUG oslo_concurrency.lockutils [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 822.254948] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2a606695-cc75-465b-9418-b73ae4d5e416 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.259983] env[63088]: DEBUG oslo_vmware.api [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Waiting for the task: (returnval){ [ 822.259983] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52d6f8b2-ac9f-ea3f-ad22-7af397fbd3f4" [ 822.259983] env[63088]: _type = "Task" [ 822.259983] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.261123] env[63088]: DEBUG oslo_concurrency.lockutils [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 822.275198] env[63088]: DEBUG oslo_vmware.api [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52d6f8b2-ac9f-ea3f-ad22-7af397fbd3f4, 'name': SearchDatastore_Task, 'duration_secs': 0.010606} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.275549] env[63088]: DEBUG oslo_concurrency.lockutils [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 822.275614] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] [instance: 4f6e2d99-7300-42e6-8abf-362ddd4e8c6d] Processing image 1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 822.275828] env[63088]: DEBUG oslo_concurrency.lockutils [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 822.275985] env[63088]: DEBUG oslo_concurrency.lockutils [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 822.276178] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 822.276420] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f3abb0af-7ff7-4b7d-8a5c-71975a8e93dc {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.284851] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 822.285098] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 822.287655] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-98ad5872-8a88-40c4-9464-72d262e23aa1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.299168] env[63088]: DEBUG oslo_vmware.api [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Waiting for the task: (returnval){ [ 822.299168] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52a4f38b-e2d9-fecc-b8d6-ed0d03f83e92" [ 822.299168] env[63088]: _type = "Task" [ 822.299168] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.305538] env[63088]: DEBUG oslo_vmware.api [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52a4f38b-e2d9-fecc-b8d6-ed0d03f83e92, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.469927] env[63088]: DEBUG nova.network.neutron [req-4bec4a38-204a-48d3-b7b8-9d362d3c27b2 req-ad4dbd71-0f93-4360-a794-9052ff60ff7e service nova] [instance: 4f6e2d99-7300-42e6-8abf-362ddd4e8c6d] Updated VIF entry in instance network info cache for port 07c78e97-0b3c-45a7-859f-bbf74ce3c5f6. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 822.470292] env[63088]: DEBUG nova.network.neutron [req-4bec4a38-204a-48d3-b7b8-9d362d3c27b2 req-ad4dbd71-0f93-4360-a794-9052ff60ff7e service nova] [instance: 4f6e2d99-7300-42e6-8abf-362ddd4e8c6d] Updating instance_info_cache with network_info: [{"id": "07c78e97-0b3c-45a7-859f-bbf74ce3c5f6", "address": "fa:16:3e:8f:bb:e9", "network": {"id": "5990b541-319a-48bf-a47f-70452bdab9ce", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-1642910243-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "960b3ebce4b747dfa84c1659118c78f0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bec1528b-3e87-477b-8ab2-02696ad47e66", "external-id": "nsx-vlan-transportzone-180", "segmentation_id": 180, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap07c78e97-0b", "ovs_interfaceid": "07c78e97-0b3c-45a7-859f-bbf74ce3c5f6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 822.479017] env[63088]: DEBUG oslo_vmware.api [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1284877, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.214031} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.479017] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: e845f60b-004f-4a5c-813e-0c183e99eba6] Extended root virtual disk {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 822.479655] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b150b3ea-939b-4307-ae87-9d4fef9da7f0 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.505674] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: e845f60b-004f-4a5c-813e-0c183e99eba6] Reconfiguring VM instance instance-00000038 to attach disk [datastore2] e845f60b-004f-4a5c-813e-0c183e99eba6/e845f60b-004f-4a5c-813e-0c183e99eba6.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 822.508859] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5bddb3ca-f762-442f-b6cf-7ee47b91b5d3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.529721] env[63088]: DEBUG oslo_vmware.api [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Waiting for the task: (returnval){ [ 822.529721] env[63088]: value = "task-1284880" [ 822.529721] env[63088]: _type = "Task" [ 822.529721] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.536406] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f127493-0034-4411-a4cd-bc6b8d95c4f8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.542098] env[63088]: DEBUG oslo_vmware.api [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1284880, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.546450] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23e49a98-92bd-43a6-bd4d-3a54a76b22f2 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.575682] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18518ac1-69ba-4d3e-acbd-f805b3722ca0 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.582944] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98392caf-25c4-4bd6-8132-1b3340805b74 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.597239] env[63088]: DEBUG nova.compute.provider_tree [None req-b5c531a2-a229-4998-9a9b-9ebb1e7afff2 tempest-ServerGroupTestJSON-1534054659 tempest-ServerGroupTestJSON-1534054659-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 822.611848] env[63088]: DEBUG oslo_vmware.api [None req-69612416-d1eb-4a97-8ff2-4c55509ccebf tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': task-1284878, 'name': PowerOnVM_Task, 'duration_secs': 0.434905} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.612139] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-69612416-d1eb-4a97-8ff2-4c55509ccebf tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: db032da3-d334-486f-a428-af1c8a3c360a] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 822.614483] env[63088]: DEBUG nova.compute.manager [None req-69612416-d1eb-4a97-8ff2-4c55509ccebf tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: db032da3-d334-486f-a428-af1c8a3c360a] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 822.615289] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75aa31a5-12b5-4664-ae9b-d235e8e9bb84 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.672831] env[63088]: DEBUG oslo_vmware.api [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1284879, 'name': Rename_Task, 'duration_secs': 0.327299} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.673106] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 822.673355] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d9d6d18c-11f7-49ab-94d5-6e00f6e4b8a7 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.680019] env[63088]: DEBUG oslo_vmware.api [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Waiting for the task: (returnval){ [ 822.680019] env[63088]: value = "task-1284881" [ 822.680019] env[63088]: _type = "Task" [ 822.680019] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.687562] env[63088]: DEBUG oslo_vmware.api [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1284881, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.706595] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Acquiring lock "refresh_cache-07fd3e6f-e0af-467c-9039-238be5c58d25" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 822.706779] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Acquired lock "refresh_cache-07fd3e6f-e0af-467c-9039-238be5c58d25" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 822.706860] env[63088]: DEBUG nova.network.neutron [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 822.738648] env[63088]: DEBUG nova.compute.manager [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] [instance: b159649b-ebd5-4c7d-9074-dc0e6395a947] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 822.762575] env[63088]: DEBUG nova.compute.manager [req-5fb0713c-3882-467e-a483-161fcaa9a875 req-1652fbcc-f923-4438-b657-3787f051efc9 service nova] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Received event network-vif-plugged-0c182796-7fb6-40ab-ba27-a43f0700dee1 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 822.762763] env[63088]: DEBUG oslo_concurrency.lockutils [req-5fb0713c-3882-467e-a483-161fcaa9a875 req-1652fbcc-f923-4438-b657-3787f051efc9 service nova] Acquiring lock "07fd3e6f-e0af-467c-9039-238be5c58d25-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 822.762902] env[63088]: DEBUG oslo_concurrency.lockutils [req-5fb0713c-3882-467e-a483-161fcaa9a875 req-1652fbcc-f923-4438-b657-3787f051efc9 service nova] Lock "07fd3e6f-e0af-467c-9039-238be5c58d25-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 822.763077] env[63088]: DEBUG oslo_concurrency.lockutils [req-5fb0713c-3882-467e-a483-161fcaa9a875 req-1652fbcc-f923-4438-b657-3787f051efc9 service nova] Lock "07fd3e6f-e0af-467c-9039-238be5c58d25-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 822.763267] env[63088]: DEBUG nova.compute.manager [req-5fb0713c-3882-467e-a483-161fcaa9a875 req-1652fbcc-f923-4438-b657-3787f051efc9 service nova] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] No waiting events found dispatching network-vif-plugged-0c182796-7fb6-40ab-ba27-a43f0700dee1 {{(pid=63088) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 822.763408] env[63088]: WARNING nova.compute.manager [req-5fb0713c-3882-467e-a483-161fcaa9a875 req-1652fbcc-f923-4438-b657-3787f051efc9 service nova] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Received unexpected event network-vif-plugged-0c182796-7fb6-40ab-ba27-a43f0700dee1 for instance with vm_state building and task_state spawning. [ 822.763568] env[63088]: DEBUG nova.compute.manager [req-5fb0713c-3882-467e-a483-161fcaa9a875 req-1652fbcc-f923-4438-b657-3787f051efc9 service nova] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Received event network-changed-0c182796-7fb6-40ab-ba27-a43f0700dee1 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 822.763789] env[63088]: DEBUG nova.compute.manager [req-5fb0713c-3882-467e-a483-161fcaa9a875 req-1652fbcc-f923-4438-b657-3787f051efc9 service nova] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Refreshing instance network info cache due to event network-changed-0c182796-7fb6-40ab-ba27-a43f0700dee1. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 822.763959] env[63088]: DEBUG oslo_concurrency.lockutils [req-5fb0713c-3882-467e-a483-161fcaa9a875 req-1652fbcc-f923-4438-b657-3787f051efc9 service nova] Acquiring lock "refresh_cache-07fd3e6f-e0af-467c-9039-238be5c58d25" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 822.808305] env[63088]: DEBUG oslo_vmware.api [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52a4f38b-e2d9-fecc-b8d6-ed0d03f83e92, 'name': SearchDatastore_Task, 'duration_secs': 0.00978} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.809145] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c9eecd46-147e-4294-9b2e-e72810774c44 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.814537] env[63088]: DEBUG oslo_vmware.api [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Waiting for the task: (returnval){ [ 822.814537] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]522a89da-5e00-9be9-7af5-03c7f017741e" [ 822.814537] env[63088]: _type = "Task" [ 822.814537] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.822048] env[63088]: DEBUG oslo_vmware.api [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]522a89da-5e00-9be9-7af5-03c7f017741e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.976071] env[63088]: DEBUG oslo_concurrency.lockutils [req-4bec4a38-204a-48d3-b7b8-9d362d3c27b2 req-ad4dbd71-0f93-4360-a794-9052ff60ff7e service nova] Releasing lock "refresh_cache-4f6e2d99-7300-42e6-8abf-362ddd4e8c6d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 823.043963] env[63088]: DEBUG oslo_vmware.api [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1284880, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.100501] env[63088]: DEBUG nova.scheduler.client.report [None req-b5c531a2-a229-4998-9a9b-9ebb1e7afff2 tempest-ServerGroupTestJSON-1534054659 tempest-ServerGroupTestJSON-1534054659-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 823.190738] env[63088]: DEBUG oslo_vmware.api [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1284881, 'name': PowerOnVM_Task} progress is 1%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.241830] env[63088]: DEBUG nova.network.neutron [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 823.325541] env[63088]: DEBUG oslo_vmware.api [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]522a89da-5e00-9be9-7af5-03c7f017741e, 'name': SearchDatastore_Task, 'duration_secs': 0.01104} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.328368] env[63088]: DEBUG oslo_concurrency.lockutils [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 823.328655] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore2] 4f6e2d99-7300-42e6-8abf-362ddd4e8c6d/4f6e2d99-7300-42e6-8abf-362ddd4e8c6d.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 823.328938] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-eb1580ea-349f-4660-9a80-64e4f9edf9f3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.336279] env[63088]: DEBUG oslo_vmware.api [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Waiting for the task: (returnval){ [ 823.336279] env[63088]: value = "task-1284882" [ 823.336279] env[63088]: _type = "Task" [ 823.336279] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.346072] env[63088]: DEBUG oslo_vmware.api [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Task: {'id': task-1284882, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.434324] env[63088]: DEBUG nova.network.neutron [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Updating instance_info_cache with network_info: [{"id": "0c182796-7fb6-40ab-ba27-a43f0700dee1", "address": "fa:16:3e:63:25:2a", "network": {"id": "de116097-3737-4dcf-972c-88c49fa8995d", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-580707267-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "937eba1a40c84534b279d5798e234625", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3a80436-f7a9-431a-acec-aca3d76e3f9b", "external-id": "cl2-zone-339", "segmentation_id": 339, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c182796-7f", "ovs_interfaceid": "0c182796-7fb6-40ab-ba27-a43f0700dee1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 823.447151] env[63088]: DEBUG nova.compute.manager [req-e14c42e1-be66-4bf8-9cc9-30fa1b4269f1 req-4cb15267-f064-4db9-a776-cecc75b90e08 service nova] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Received event network-changed-d5a10951-ae2e-48f5-b0ca-b1f144e5fe3a {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 823.447427] env[63088]: DEBUG nova.compute.manager [req-e14c42e1-be66-4bf8-9cc9-30fa1b4269f1 req-4cb15267-f064-4db9-a776-cecc75b90e08 service nova] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Refreshing instance network info cache due to event network-changed-d5a10951-ae2e-48f5-b0ca-b1f144e5fe3a. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 823.447812] env[63088]: DEBUG oslo_concurrency.lockutils [req-e14c42e1-be66-4bf8-9cc9-30fa1b4269f1 req-4cb15267-f064-4db9-a776-cecc75b90e08 service nova] Acquiring lock "refresh_cache-4971b24c-6710-4f50-9846-727dad264b1f" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 823.448021] env[63088]: DEBUG oslo_concurrency.lockutils [req-e14c42e1-be66-4bf8-9cc9-30fa1b4269f1 req-4cb15267-f064-4db9-a776-cecc75b90e08 service nova] Acquired lock "refresh_cache-4971b24c-6710-4f50-9846-727dad264b1f" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.448284] env[63088]: DEBUG nova.network.neutron [req-e14c42e1-be66-4bf8-9cc9-30fa1b4269f1 req-4cb15267-f064-4db9-a776-cecc75b90e08 service nova] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Refreshing network info cache for port d5a10951-ae2e-48f5-b0ca-b1f144e5fe3a {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 823.540459] env[63088]: DEBUG oslo_vmware.api [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1284880, 'name': ReconfigVM_Task, 'duration_secs': 0.687647} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.540744] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: e845f60b-004f-4a5c-813e-0c183e99eba6] Reconfigured VM instance instance-00000038 to attach disk [datastore2] e845f60b-004f-4a5c-813e-0c183e99eba6/e845f60b-004f-4a5c-813e-0c183e99eba6.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 823.541403] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f5040c53-a4d3-43d6-89c4-2bc2681a0c75 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.548303] env[63088]: DEBUG oslo_vmware.api [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Waiting for the task: (returnval){ [ 823.548303] env[63088]: value = "task-1284883" [ 823.548303] env[63088]: _type = "Task" [ 823.548303] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.558029] env[63088]: DEBUG oslo_vmware.api [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1284883, 'name': Rename_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.606271] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b5c531a2-a229-4998-9a9b-9ebb1e7afff2 tempest-ServerGroupTestJSON-1534054659 tempest-ServerGroupTestJSON-1534054659-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.881s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 823.606971] env[63088]: ERROR nova.compute.manager [None req-b5c531a2-a229-4998-9a9b-9ebb1e7afff2 tempest-ServerGroupTestJSON-1534054659 tempest-ServerGroupTestJSON-1534054659-project-member] [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3235d186-bfbf-41f3-88af-330b3981042d, please check neutron logs for more information. [ 823.606971] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] Traceback (most recent call last): [ 823.606971] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 823.606971] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] self.driver.spawn(context, instance, image_meta, [ 823.606971] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 823.606971] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 823.606971] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 823.606971] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] vm_ref = self.build_virtual_machine(instance, [ 823.606971] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 823.606971] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] vif_infos = vmwarevif.get_vif_info(self._session, [ 823.606971] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 823.607330] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] for vif in network_info: [ 823.607330] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 823.607330] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] return self._sync_wrapper(fn, *args, **kwargs) [ 823.607330] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 823.607330] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] self.wait() [ 823.607330] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 823.607330] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] self[:] = self._gt.wait() [ 823.607330] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 823.607330] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] return self._exit_event.wait() [ 823.607330] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 823.607330] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] result = hub.switch() [ 823.607330] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 823.607330] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] return self.greenlet.switch() [ 823.607683] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 823.607683] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] result = function(*args, **kwargs) [ 823.607683] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 823.607683] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] return func(*args, **kwargs) [ 823.607683] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 823.607683] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] raise e [ 823.607683] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 823.607683] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] nwinfo = self.network_api.allocate_for_instance( [ 823.607683] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 823.607683] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] created_port_ids = self._update_ports_for_instance( [ 823.607683] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 823.607683] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] with excutils.save_and_reraise_exception(): [ 823.607683] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 823.608045] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] self.force_reraise() [ 823.608045] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 823.608045] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] raise self.value [ 823.608045] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 823.608045] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] updated_port = self._update_port( [ 823.608045] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 823.608045] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] _ensure_no_port_binding_failure(port) [ 823.608045] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 823.608045] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] raise exception.PortBindingFailed(port_id=port['id']) [ 823.608045] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] nova.exception.PortBindingFailed: Binding failed for port 3235d186-bfbf-41f3-88af-330b3981042d, please check neutron logs for more information. [ 823.608045] env[63088]: ERROR nova.compute.manager [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] [ 823.608331] env[63088]: DEBUG nova.compute.utils [None req-b5c531a2-a229-4998-9a9b-9ebb1e7afff2 tempest-ServerGroupTestJSON-1534054659 tempest-ServerGroupTestJSON-1534054659-project-member] [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] Binding failed for port 3235d186-bfbf-41f3-88af-330b3981042d, please check neutron logs for more information. {{(pid=63088) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 823.609081] env[63088]: DEBUG oslo_concurrency.lockutils [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.317s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 823.610770] env[63088]: INFO nova.compute.claims [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] [instance: 34ce411f-40c8-446e-b685-cd3ce07663dd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 823.614093] env[63088]: DEBUG nova.compute.manager [None req-b5c531a2-a229-4998-9a9b-9ebb1e7afff2 tempest-ServerGroupTestJSON-1534054659 tempest-ServerGroupTestJSON-1534054659-project-member] [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] Build of instance 53ea5c01-4be2-44a8-b74f-0d74031dfdd7 was re-scheduled: Binding failed for port 3235d186-bfbf-41f3-88af-330b3981042d, please check neutron logs for more information. {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 823.614206] env[63088]: DEBUG nova.compute.manager [None req-b5c531a2-a229-4998-9a9b-9ebb1e7afff2 tempest-ServerGroupTestJSON-1534054659 tempest-ServerGroupTestJSON-1534054659-project-member] [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] Unplugging VIFs for instance {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 823.614454] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b5c531a2-a229-4998-9a9b-9ebb1e7afff2 tempest-ServerGroupTestJSON-1534054659 tempest-ServerGroupTestJSON-1534054659-project-member] Acquiring lock "refresh_cache-53ea5c01-4be2-44a8-b74f-0d74031dfdd7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 823.614617] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b5c531a2-a229-4998-9a9b-9ebb1e7afff2 tempest-ServerGroupTestJSON-1534054659 tempest-ServerGroupTestJSON-1534054659-project-member] Acquired lock "refresh_cache-53ea5c01-4be2-44a8-b74f-0d74031dfdd7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.614782] env[63088]: DEBUG nova.network.neutron [None req-b5c531a2-a229-4998-9a9b-9ebb1e7afff2 tempest-ServerGroupTestJSON-1534054659 tempest-ServerGroupTestJSON-1534054659-project-member] [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 823.691725] env[63088]: DEBUG oslo_vmware.api [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1284881, 'name': PowerOnVM_Task} progress is 64%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.752021] env[63088]: DEBUG nova.compute.manager [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] [instance: b159649b-ebd5-4c7d-9074-dc0e6395a947] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 823.775225] env[63088]: DEBUG nova.virt.hardware [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 823.775484] env[63088]: DEBUG nova.virt.hardware [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 823.775637] env[63088]: DEBUG nova.virt.hardware [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 823.775818] env[63088]: DEBUG nova.virt.hardware [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 823.775964] env[63088]: DEBUG nova.virt.hardware [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 823.776123] env[63088]: DEBUG nova.virt.hardware [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 823.776361] env[63088]: DEBUG nova.virt.hardware [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 823.776534] env[63088]: DEBUG nova.virt.hardware [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 823.776703] env[63088]: DEBUG nova.virt.hardware [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 823.776868] env[63088]: DEBUG nova.virt.hardware [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 823.777057] env[63088]: DEBUG nova.virt.hardware [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 823.777995] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab155550-726f-4f60-b1c8-73f0e148b244 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.787455] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-caae6264-e8b8-4d13-bd0c-b4eee28d3a4d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.803974] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] [instance: b159649b-ebd5-4c7d-9074-dc0e6395a947] Instance VIF info [] {{(pid=63088) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 823.809979] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Creating folder: Project (e55c4d9cc1c24ec4947047f4916471e2). Parent ref: group-v275816. {{(pid=63088) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 823.810832] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-323e3520-2185-4ead-8ff3-9dfeb452b96c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.821800] env[63088]: INFO nova.virt.vmwareapi.vm_util [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Created folder: Project (e55c4d9cc1c24ec4947047f4916471e2) in parent group-v275816. [ 823.822026] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Creating folder: Instances. Parent ref: group-v275851. {{(pid=63088) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 823.822291] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9557980c-6a97-47ef-bc8a-65a33f435458 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.832342] env[63088]: INFO nova.virt.vmwareapi.vm_util [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Created folder: Instances in parent group-v275851. [ 823.832685] env[63088]: DEBUG oslo.service.loopingcall [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 823.832904] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b159649b-ebd5-4c7d-9074-dc0e6395a947] Creating VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 823.833142] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5dacd9a5-58ba-4c3f-a47e-34265267ec33 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.854972] env[63088]: DEBUG oslo_vmware.api [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Task: {'id': task-1284882, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.856420] env[63088]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 823.856420] env[63088]: value = "task-1284886" [ 823.856420] env[63088]: _type = "Task" [ 823.856420] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.865338] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1284886, 'name': CreateVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.878958] env[63088]: DEBUG oslo_concurrency.lockutils [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Acquiring lock "58d561fe-26aa-4e94-8d55-cc70c361b479" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 823.879261] env[63088]: DEBUG oslo_concurrency.lockutils [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Lock "58d561fe-26aa-4e94-8d55-cc70c361b479" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 823.937867] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Releasing lock "refresh_cache-07fd3e6f-e0af-467c-9039-238be5c58d25" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 823.938232] env[63088]: DEBUG nova.compute.manager [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Instance network_info: |[{"id": "0c182796-7fb6-40ab-ba27-a43f0700dee1", "address": "fa:16:3e:63:25:2a", "network": {"id": "de116097-3737-4dcf-972c-88c49fa8995d", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-580707267-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "937eba1a40c84534b279d5798e234625", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3a80436-f7a9-431a-acec-aca3d76e3f9b", "external-id": "cl2-zone-339", "segmentation_id": 339, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c182796-7f", "ovs_interfaceid": "0c182796-7fb6-40ab-ba27-a43f0700dee1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 823.938601] env[63088]: DEBUG oslo_concurrency.lockutils [req-5fb0713c-3882-467e-a483-161fcaa9a875 req-1652fbcc-f923-4438-b657-3787f051efc9 service nova] Acquired lock "refresh_cache-07fd3e6f-e0af-467c-9039-238be5c58d25" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.938834] env[63088]: DEBUG nova.network.neutron [req-5fb0713c-3882-467e-a483-161fcaa9a875 req-1652fbcc-f923-4438-b657-3787f051efc9 service nova] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Refreshing network info cache for port 0c182796-7fb6-40ab-ba27-a43f0700dee1 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 823.940254] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:63:25:2a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f3a80436-f7a9-431a-acec-aca3d76e3f9b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0c182796-7fb6-40ab-ba27-a43f0700dee1', 'vif_model': 'vmxnet3'}] {{(pid=63088) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 823.948088] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Creating folder: Project (937eba1a40c84534b279d5798e234625). Parent ref: group-v275816. {{(pid=63088) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 823.949445] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-574ef39e-c51f-420d-a73a-81488531a277 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.961147] env[63088]: INFO nova.virt.vmwareapi.vm_util [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Created folder: Project (937eba1a40c84534b279d5798e234625) in parent group-v275816. [ 823.961343] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Creating folder: Instances. Parent ref: group-v275854. {{(pid=63088) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 823.961575] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5aaa6b79-908c-46d7-bc5a-1006585b86f6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.970117] env[63088]: INFO nova.virt.vmwareapi.vm_util [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Created folder: Instances in parent group-v275854. [ 823.970349] env[63088]: DEBUG oslo.service.loopingcall [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 823.970539] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Creating VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 823.970779] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b60ae743-cfad-4f33-b9d1-e7eb28075ec7 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.990785] env[63088]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 823.990785] env[63088]: value = "task-1284889" [ 823.990785] env[63088]: _type = "Task" [ 823.990785] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.000207] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1284889, 'name': CreateVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.058577] env[63088]: DEBUG oslo_vmware.api [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1284883, 'name': Rename_Task} progress is 99%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.194224] env[63088]: DEBUG oslo_vmware.api [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1284881, 'name': PowerOnVM_Task, 'duration_secs': 1.40165} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.194550] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 824.194771] env[63088]: INFO nova.compute.manager [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] Took 9.50 seconds to spawn the instance on the hypervisor. [ 824.195312] env[63088]: DEBUG nova.compute.manager [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 824.195820] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08c79a14-1ef9-438a-ab80-1a3ebf7a436c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.224614] env[63088]: DEBUG nova.network.neutron [None req-b5c531a2-a229-4998-9a9b-9ebb1e7afff2 tempest-ServerGroupTestJSON-1534054659 tempest-ServerGroupTestJSON-1534054659-project-member] [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 824.356296] env[63088]: DEBUG oslo_vmware.api [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Task: {'id': task-1284882, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.575565} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.356652] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore2] 4f6e2d99-7300-42e6-8abf-362ddd4e8c6d/4f6e2d99-7300-42e6-8abf-362ddd4e8c6d.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 824.356957] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] [instance: 4f6e2d99-7300-42e6-8abf-362ddd4e8c6d] Extending root virtual disk to 1048576 {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 824.357227] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7119a965-f591-4f06-9216-b5972fca078f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.368588] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1284886, 'name': CreateVM_Task} progress is 25%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.370457] env[63088]: DEBUG oslo_vmware.api [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Waiting for the task: (returnval){ [ 824.370457] env[63088]: value = "task-1284890" [ 824.370457] env[63088]: _type = "Task" [ 824.370457] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.374528] env[63088]: DEBUG nova.network.neutron [None req-b5c531a2-a229-4998-9a9b-9ebb1e7afff2 tempest-ServerGroupTestJSON-1534054659 tempest-ServerGroupTestJSON-1534054659-project-member] [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 824.378933] env[63088]: DEBUG oslo_vmware.api [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Task: {'id': task-1284890, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.450652] env[63088]: DEBUG nova.network.neutron [req-e14c42e1-be66-4bf8-9cc9-30fa1b4269f1 req-4cb15267-f064-4db9-a776-cecc75b90e08 service nova] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Updated VIF entry in instance network info cache for port d5a10951-ae2e-48f5-b0ca-b1f144e5fe3a. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 824.451111] env[63088]: DEBUG nova.network.neutron [req-e14c42e1-be66-4bf8-9cc9-30fa1b4269f1 req-4cb15267-f064-4db9-a776-cecc75b90e08 service nova] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Updating instance_info_cache with network_info: [{"id": "d5a10951-ae2e-48f5-b0ca-b1f144e5fe3a", "address": "fa:16:3e:95:96:0c", "network": {"id": "1174a284-d3ca-4f9e-aa81-13ee9a693e55", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1994276040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.166", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa3d24a1a6c0430985fd80365d986ee1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd5a10951-ae", "ovs_interfaceid": "d5a10951-ae2e-48f5-b0ca-b1f144e5fe3a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 824.502572] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1284889, 'name': CreateVM_Task} progress is 25%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.562034] env[63088]: DEBUG oslo_vmware.api [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1284883, 'name': Rename_Task} progress is 99%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.694214] env[63088]: DEBUG nova.network.neutron [req-5fb0713c-3882-467e-a483-161fcaa9a875 req-1652fbcc-f923-4438-b657-3787f051efc9 service nova] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Updated VIF entry in instance network info cache for port 0c182796-7fb6-40ab-ba27-a43f0700dee1. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 824.694571] env[63088]: DEBUG nova.network.neutron [req-5fb0713c-3882-467e-a483-161fcaa9a875 req-1652fbcc-f923-4438-b657-3787f051efc9 service nova] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Updating instance_info_cache with network_info: [{"id": "0c182796-7fb6-40ab-ba27-a43f0700dee1", "address": "fa:16:3e:63:25:2a", "network": {"id": "de116097-3737-4dcf-972c-88c49fa8995d", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-580707267-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "937eba1a40c84534b279d5798e234625", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3a80436-f7a9-431a-acec-aca3d76e3f9b", "external-id": "cl2-zone-339", "segmentation_id": 339, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c182796-7f", "ovs_interfaceid": "0c182796-7fb6-40ab-ba27-a43f0700dee1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 824.714032] env[63088]: INFO nova.compute.manager [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] Took 38.87 seconds to build instance. [ 824.868281] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1284886, 'name': CreateVM_Task, 'duration_secs': 0.734212} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.870970] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b159649b-ebd5-4c7d-9074-dc0e6395a947] Created VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 824.871866] env[63088]: DEBUG oslo_concurrency.lockutils [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 824.872086] env[63088]: DEBUG oslo_concurrency.lockutils [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 824.872365] env[63088]: DEBUG oslo_concurrency.lockutils [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 824.875730] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ac246a63-66c9-4077-83bf-74459908a477 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.882241] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b5c531a2-a229-4998-9a9b-9ebb1e7afff2 tempest-ServerGroupTestJSON-1534054659 tempest-ServerGroupTestJSON-1534054659-project-member] Releasing lock "refresh_cache-53ea5c01-4be2-44a8-b74f-0d74031dfdd7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 824.882461] env[63088]: DEBUG nova.compute.manager [None req-b5c531a2-a229-4998-9a9b-9ebb1e7afff2 tempest-ServerGroupTestJSON-1534054659 tempest-ServerGroupTestJSON-1534054659-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63088) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 824.882643] env[63088]: DEBUG nova.compute.manager [None req-b5c531a2-a229-4998-9a9b-9ebb1e7afff2 tempest-ServerGroupTestJSON-1534054659 tempest-ServerGroupTestJSON-1534054659-project-member] [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 824.882810] env[63088]: DEBUG nova.network.neutron [None req-b5c531a2-a229-4998-9a9b-9ebb1e7afff2 tempest-ServerGroupTestJSON-1534054659 tempest-ServerGroupTestJSON-1534054659-project-member] [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 824.884436] env[63088]: DEBUG oslo_vmware.api [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Task: {'id': task-1284890, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065713} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.887850] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] [instance: 4f6e2d99-7300-42e6-8abf-362ddd4e8c6d] Extended root virtual disk {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 824.888237] env[63088]: DEBUG oslo_vmware.api [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Waiting for the task: (returnval){ [ 824.888237] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]528f21be-0357-501d-5ec9-2487a74db76d" [ 824.888237] env[63088]: _type = "Task" [ 824.888237] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.889420] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37a6e925-0345-4ef0-8fa2-c4010d5814f8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.913986] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] [instance: 4f6e2d99-7300-42e6-8abf-362ddd4e8c6d] Reconfiguring VM instance instance-00000039 to attach disk [datastore2] 4f6e2d99-7300-42e6-8abf-362ddd4e8c6d/4f6e2d99-7300-42e6-8abf-362ddd4e8c6d.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 824.918488] env[63088]: DEBUG nova.network.neutron [None req-b5c531a2-a229-4998-9a9b-9ebb1e7afff2 tempest-ServerGroupTestJSON-1534054659 tempest-ServerGroupTestJSON-1534054659-project-member] [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 824.920049] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-88f45608-bbf3-4927-aeb8-38491d876f18 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.934874] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec394165-7169-4cd3-93c7-8df249b6f93c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.937918] env[63088]: DEBUG oslo_vmware.api [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]528f21be-0357-501d-5ec9-2487a74db76d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.938988] env[63088]: DEBUG nova.network.neutron [None req-b5c531a2-a229-4998-9a9b-9ebb1e7afff2 tempest-ServerGroupTestJSON-1534054659 tempest-ServerGroupTestJSON-1534054659-project-member] [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 824.947232] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49695126-7197-4493-b4b1-a501c227cd6b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.950502] env[63088]: DEBUG oslo_vmware.api [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Waiting for the task: (returnval){ [ 824.950502] env[63088]: value = "task-1284891" [ 824.950502] env[63088]: _type = "Task" [ 824.950502] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.978694] env[63088]: DEBUG oslo_concurrency.lockutils [req-e14c42e1-be66-4bf8-9cc9-30fa1b4269f1 req-4cb15267-f064-4db9-a776-cecc75b90e08 service nova] Releasing lock "refresh_cache-4971b24c-6710-4f50-9846-727dad264b1f" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 824.980319] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55d333e1-9edc-4e85-b6d1-423a3042a36e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.986117] env[63088]: DEBUG oslo_vmware.api [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Task: {'id': task-1284891, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.990908] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a921c4b-7cd8-43f9-b1c3-91e5ad43e461 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.011716] env[63088]: DEBUG nova.compute.provider_tree [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 825.012740] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1284889, 'name': CreateVM_Task} progress is 99%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.059956] env[63088]: DEBUG oslo_vmware.api [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1284883, 'name': Rename_Task, 'duration_secs': 1.254753} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.060374] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: e845f60b-004f-4a5c-813e-0c183e99eba6] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 825.060772] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f382b808-5784-4be4-a88b-3dbef845f20f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.067121] env[63088]: DEBUG oslo_vmware.api [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Waiting for the task: (returnval){ [ 825.067121] env[63088]: value = "task-1284892" [ 825.067121] env[63088]: _type = "Task" [ 825.067121] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.074477] env[63088]: DEBUG oslo_vmware.api [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1284892, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.198243] env[63088]: DEBUG oslo_concurrency.lockutils [req-5fb0713c-3882-467e-a483-161fcaa9a875 req-1652fbcc-f923-4438-b657-3787f051efc9 service nova] Releasing lock "refresh_cache-07fd3e6f-e0af-467c-9039-238be5c58d25" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 825.216938] env[63088]: DEBUG oslo_concurrency.lockutils [None req-dd07d347-13c9-49bb-b70b-239f1ac18f2e tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Lock "4847f15e-98d4-401c-91ff-067e84a85727" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 81.400s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 825.401866] env[63088]: DEBUG oslo_vmware.api [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]528f21be-0357-501d-5ec9-2487a74db76d, 'name': SearchDatastore_Task, 'duration_secs': 0.018426} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.402252] env[63088]: DEBUG oslo_concurrency.lockutils [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 825.402512] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] [instance: b159649b-ebd5-4c7d-9074-dc0e6395a947] Processing image 1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 825.402790] env[63088]: DEBUG oslo_concurrency.lockutils [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 825.402949] env[63088]: DEBUG oslo_concurrency.lockutils [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 825.403143] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 825.403400] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ba63c793-5e70-4543-b59b-00207b6ef35f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.412076] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 825.412284] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 825.413009] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b393ee74-97c8-43b3-bfa7-944e9571bf2d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.418180] env[63088]: DEBUG oslo_vmware.api [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Waiting for the task: (returnval){ [ 825.418180] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52969b05-35e4-4f72-d12e-9cd4fcbfa28b" [ 825.418180] env[63088]: _type = "Task" [ 825.418180] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.425635] env[63088]: DEBUG oslo_vmware.api [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52969b05-35e4-4f72-d12e-9cd4fcbfa28b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.441425] env[63088]: INFO nova.compute.manager [None req-b5c531a2-a229-4998-9a9b-9ebb1e7afff2 tempest-ServerGroupTestJSON-1534054659 tempest-ServerGroupTestJSON-1534054659-project-member] [instance: 53ea5c01-4be2-44a8-b74f-0d74031dfdd7] Took 0.56 seconds to deallocate network for instance. [ 825.460676] env[63088]: DEBUG oslo_vmware.api [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Task: {'id': task-1284891, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.508104] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1284889, 'name': CreateVM_Task} progress is 99%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.515719] env[63088]: DEBUG nova.scheduler.client.report [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 825.578344] env[63088]: DEBUG oslo_vmware.api [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1284892, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.722026] env[63088]: DEBUG nova.compute.manager [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 7f341d80-450a-47b8-b26d-15f2ce3e378c] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 825.930057] env[63088]: DEBUG oslo_vmware.api [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52969b05-35e4-4f72-d12e-9cd4fcbfa28b, 'name': SearchDatastore_Task, 'duration_secs': 0.014356} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.930949] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-92f73e8a-1e42-4a8c-8ca4-81363efdc4bd {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.936936] env[63088]: DEBUG oslo_vmware.api [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Waiting for the task: (returnval){ [ 825.936936] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52453864-1022-ec86-2f0d-f801b51369a3" [ 825.936936] env[63088]: _type = "Task" [ 825.936936] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.948955] env[63088]: DEBUG oslo_vmware.api [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52453864-1022-ec86-2f0d-f801b51369a3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.960160] env[63088]: DEBUG oslo_vmware.api [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Task: {'id': task-1284891, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.001406] env[63088]: INFO nova.compute.manager [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] Rebuilding instance [ 826.008452] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1284889, 'name': CreateVM_Task, 'duration_secs': 1.651645} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.008614] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Created VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 826.009270] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 826.009436] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 826.009800] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 826.010038] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eeaa7f80-2598-4da6-8ce1-53fded5fb75a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.014215] env[63088]: DEBUG oslo_vmware.api [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Waiting for the task: (returnval){ [ 826.014215] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52b57f8d-7d53-1ab4-ac16-548fa264e632" [ 826.014215] env[63088]: _type = "Task" [ 826.014215] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.020884] env[63088]: DEBUG oslo_concurrency.lockutils [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.412s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 826.021156] env[63088]: DEBUG nova.compute.manager [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] [instance: 34ce411f-40c8-446e-b685-cd3ce07663dd] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 826.026865] env[63088]: DEBUG oslo_concurrency.lockutils [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.177s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 826.028337] env[63088]: INFO nova.compute.claims [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] [instance: 1d421bf0-925d-408c-a612-b659da92799f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 826.031423] env[63088]: DEBUG oslo_vmware.api [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52b57f8d-7d53-1ab4-ac16-548fa264e632, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.053119] env[63088]: DEBUG nova.compute.manager [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 826.054067] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3745271-09a9-4d62-98a3-a0472406c745 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.076565] env[63088]: DEBUG oslo_vmware.api [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1284892, 'name': PowerOnVM_Task, 'duration_secs': 0.852379} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.076767] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: e845f60b-004f-4a5c-813e-0c183e99eba6] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 826.076963] env[63088]: INFO nova.compute.manager [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: e845f60b-004f-4a5c-813e-0c183e99eba6] Took 9.01 seconds to spawn the instance on the hypervisor. [ 826.077155] env[63088]: DEBUG nova.compute.manager [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: e845f60b-004f-4a5c-813e-0c183e99eba6] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 826.077908] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7aab14d5-b89b-4747-95af-ec56d6338d22 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.238522] env[63088]: DEBUG oslo_concurrency.lockutils [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 826.447521] env[63088]: DEBUG oslo_vmware.api [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52453864-1022-ec86-2f0d-f801b51369a3, 'name': SearchDatastore_Task, 'duration_secs': 0.010293} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.447820] env[63088]: DEBUG oslo_concurrency.lockutils [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 826.448088] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] b159649b-ebd5-4c7d-9074-dc0e6395a947/b159649b-ebd5-4c7d-9074-dc0e6395a947.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 826.448345] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a20d9908-fa3a-4c57-9267-3c929944a008 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.457409] env[63088]: DEBUG oslo_vmware.api [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Waiting for the task: (returnval){ [ 826.457409] env[63088]: value = "task-1284893" [ 826.457409] env[63088]: _type = "Task" [ 826.457409] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.460296] env[63088]: DEBUG oslo_vmware.api [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Task: {'id': task-1284891, 'name': ReconfigVM_Task, 'duration_secs': 1.093723} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.463100] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] [instance: 4f6e2d99-7300-42e6-8abf-362ddd4e8c6d] Reconfigured VM instance instance-00000039 to attach disk [datastore2] 4f6e2d99-7300-42e6-8abf-362ddd4e8c6d/4f6e2d99-7300-42e6-8abf-362ddd4e8c6d.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 826.463617] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5803c246-cbbf-4736-867b-2b9f79cb55d1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.469667] env[63088]: DEBUG oslo_vmware.api [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Task: {'id': task-1284893, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.470848] env[63088]: DEBUG oslo_vmware.api [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Waiting for the task: (returnval){ [ 826.470848] env[63088]: value = "task-1284894" [ 826.470848] env[63088]: _type = "Task" [ 826.470848] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.474585] env[63088]: INFO nova.scheduler.client.report [None req-b5c531a2-a229-4998-9a9b-9ebb1e7afff2 tempest-ServerGroupTestJSON-1534054659 tempest-ServerGroupTestJSON-1534054659-project-member] Deleted allocations for instance 53ea5c01-4be2-44a8-b74f-0d74031dfdd7 [ 826.486585] env[63088]: DEBUG oslo_vmware.api [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Task: {'id': task-1284894, 'name': Rename_Task} progress is 10%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.524840] env[63088]: DEBUG oslo_vmware.api [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52b57f8d-7d53-1ab4-ac16-548fa264e632, 'name': SearchDatastore_Task, 'duration_secs': 0.009892} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.525262] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 826.525459] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Processing image 1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 826.525715] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 826.525879] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 826.526083] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 826.526371] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2727e592-6d78-4ba6-8d6a-ea8c5ae790f9 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.537314] env[63088]: DEBUG nova.compute.utils [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 826.538140] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 826.538317] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 826.539564] env[63088]: DEBUG nova.compute.manager [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] [instance: 34ce411f-40c8-446e-b685-cd3ce07663dd] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 826.539791] env[63088]: DEBUG nova.network.neutron [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] [instance: 34ce411f-40c8-446e-b685-cd3ce07663dd] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 826.541482] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c9df42f8-18ad-49b6-af37-05fe362e517f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.547712] env[63088]: DEBUG oslo_vmware.api [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Waiting for the task: (returnval){ [ 826.547712] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52b3e239-84e9-0141-9733-d3699268af65" [ 826.547712] env[63088]: _type = "Task" [ 826.547712] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.557849] env[63088]: DEBUG oslo_vmware.api [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52b3e239-84e9-0141-9733-d3699268af65, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.564493] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 826.564764] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6deab526-679c-41a2-ae61-9d22aced9c87 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.570399] env[63088]: DEBUG oslo_vmware.api [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Waiting for the task: (returnval){ [ 826.570399] env[63088]: value = "task-1284895" [ 826.570399] env[63088]: _type = "Task" [ 826.570399] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.578225] env[63088]: DEBUG oslo_vmware.api [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1284895, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.595126] env[63088]: INFO nova.compute.manager [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: e845f60b-004f-4a5c-813e-0c183e99eba6] Took 38.80 seconds to build instance. [ 826.599693] env[63088]: DEBUG nova.policy [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '673296498bef4559864184e7e9c4a41c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a5f62816dd0243da9be2073086b136d8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 826.970712] env[63088]: DEBUG oslo_vmware.api [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Task: {'id': task-1284893, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.4775} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.970977] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] b159649b-ebd5-4c7d-9074-dc0e6395a947/b159649b-ebd5-4c7d-9074-dc0e6395a947.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 826.971204] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] [instance: b159649b-ebd5-4c7d-9074-dc0e6395a947] Extending root virtual disk to 1048576 {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 826.971500] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cd5fda18-60e1-4ce9-a5a3-e467b4d77783 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.982645] env[63088]: DEBUG oslo_vmware.api [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Task: {'id': task-1284894, 'name': Rename_Task, 'duration_secs': 0.132472} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.982645] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] [instance: 4f6e2d99-7300-42e6-8abf-362ddd4e8c6d] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 826.982645] env[63088]: DEBUG oslo_vmware.api [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Waiting for the task: (returnval){ [ 826.982645] env[63088]: value = "task-1284896" [ 826.982645] env[63088]: _type = "Task" [ 826.982645] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.982645] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5349f003-48f1-4398-b615-8751b456e466 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.988220] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b5c531a2-a229-4998-9a9b-9ebb1e7afff2 tempest-ServerGroupTestJSON-1534054659 tempest-ServerGroupTestJSON-1534054659-project-member] Lock "53ea5c01-4be2-44a8-b74f-0d74031dfdd7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 133.263s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 826.993361] env[63088]: DEBUG oslo_vmware.api [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Task: {'id': task-1284896, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.994809] env[63088]: DEBUG oslo_vmware.api [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Waiting for the task: (returnval){ [ 826.994809] env[63088]: value = "task-1284897" [ 826.994809] env[63088]: _type = "Task" [ 826.994809] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.002626] env[63088]: DEBUG oslo_vmware.api [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Task: {'id': task-1284897, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.041914] env[63088]: DEBUG nova.compute.manager [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] [instance: 34ce411f-40c8-446e-b685-cd3ce07663dd] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 827.048745] env[63088]: DEBUG nova.network.neutron [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] [instance: 34ce411f-40c8-446e-b685-cd3ce07663dd] Successfully created port: 44d58f1e-f90f-4aea-b8fd-5f9c7ff195cb {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 827.061013] env[63088]: DEBUG oslo_vmware.api [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52b3e239-84e9-0141-9733-d3699268af65, 'name': SearchDatastore_Task, 'duration_secs': 0.008584} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.061870] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9c53c4a5-5878-4de9-9797-82c16ae7e558 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.067008] env[63088]: DEBUG oslo_vmware.api [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Waiting for the task: (returnval){ [ 827.067008] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52ef87f2-4e90-c16c-bf0a-2af984648c26" [ 827.067008] env[63088]: _type = "Task" [ 827.067008] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.079829] env[63088]: DEBUG oslo_vmware.api [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52ef87f2-4e90-c16c-bf0a-2af984648c26, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.083647] env[63088]: DEBUG oslo_vmware.api [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1284895, 'name': PowerOffVM_Task, 'duration_secs': 0.233437} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.083945] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 827.084161] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 827.084852] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9450c9e0-a4f9-4f12-8679-8b061046e716 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.091295] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] Unregistering the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 827.093686] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cdcc18c4-dbe3-46df-ad48-97a0283242cf {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.097395] env[63088]: DEBUG oslo_concurrency.lockutils [None req-45a914b1-8dc2-4a7f-91b5-a221c060c818 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lock "e845f60b-004f-4a5c-813e-0c183e99eba6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 80.798s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 827.155284] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] Unregistered the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 827.155558] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] Deleting contents of the VM from datastore datastore2 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 827.155777] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Deleting the datastore file [datastore2] 4847f15e-98d4-401c-91ff-067e84a85727 {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 827.156093] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c1e51cdf-a4f0-43f3-8020-2f8565358f14 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.162967] env[63088]: DEBUG oslo_vmware.api [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Waiting for the task: (returnval){ [ 827.162967] env[63088]: value = "task-1284899" [ 827.162967] env[63088]: _type = "Task" [ 827.162967] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.170661] env[63088]: DEBUG oslo_vmware.api [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1284899, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.292821] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab3a7818-acf6-4351-8b8f-1a6d0e1d4da6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.300292] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ea4b753-bf27-4051-8715-c2b9349bd4e3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.332520] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e66baf24-3d05-4a1d-ba90-fa1d6753e5a4 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.340172] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1418479a-4fbf-4084-b16b-3c2b8454e6ae {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.354485] env[63088]: DEBUG nova.compute.provider_tree [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 827.490461] env[63088]: DEBUG nova.compute.manager [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 8b564e29-96cf-4abf-963d-142b413fb464] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 827.496459] env[63088]: DEBUG oslo_vmware.api [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Task: {'id': task-1284896, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069282} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.499913] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] [instance: b159649b-ebd5-4c7d-9074-dc0e6395a947] Extended root virtual disk {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 827.500724] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78c10094-4c7b-452a-b4f1-d2ebd334ba19 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.509108] env[63088]: DEBUG oslo_vmware.api [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Task: {'id': task-1284897, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.526339] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] [instance: b159649b-ebd5-4c7d-9074-dc0e6395a947] Reconfiguring VM instance instance-0000003b to attach disk [datastore1] b159649b-ebd5-4c7d-9074-dc0e6395a947/b159649b-ebd5-4c7d-9074-dc0e6395a947.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 827.527344] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6dd5c198-b7ba-4c5e-bf8b-6915720e5c29 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.549027] env[63088]: DEBUG oslo_vmware.api [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Waiting for the task: (returnval){ [ 827.549027] env[63088]: value = "task-1284900" [ 827.549027] env[63088]: _type = "Task" [ 827.549027] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.560381] env[63088]: DEBUG oslo_vmware.api [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Task: {'id': task-1284900, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.579205] env[63088]: DEBUG oslo_vmware.api [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52ef87f2-4e90-c16c-bf0a-2af984648c26, 'name': SearchDatastore_Task, 'duration_secs': 0.017565} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.579205] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 827.579362] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore2] 07fd3e6f-e0af-467c-9039-238be5c58d25/07fd3e6f-e0af-467c-9039-238be5c58d25.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 827.579571] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-dd8bc088-8d5d-4c88-a2c5-58939832a383 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.586033] env[63088]: DEBUG oslo_vmware.api [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Waiting for the task: (returnval){ [ 827.586033] env[63088]: value = "task-1284901" [ 827.586033] env[63088]: _type = "Task" [ 827.586033] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.594859] env[63088]: DEBUG oslo_vmware.api [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Task: {'id': task-1284901, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.599623] env[63088]: DEBUG nova.compute.manager [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 827.673119] env[63088]: DEBUG oslo_vmware.api [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1284899, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.205867} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.673390] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 827.673574] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] Deleted contents of the VM from datastore datastore2 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 827.673749] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 827.860109] env[63088]: DEBUG nova.scheduler.client.report [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 827.905781] env[63088]: DEBUG oslo_concurrency.lockutils [None req-049e65c1-0c59-41b7-8448-5a3e6eb7d38f tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquiring lock "e845f60b-004f-4a5c-813e-0c183e99eba6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 827.906097] env[63088]: DEBUG oslo_concurrency.lockutils [None req-049e65c1-0c59-41b7-8448-5a3e6eb7d38f tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lock "e845f60b-004f-4a5c-813e-0c183e99eba6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 827.906309] env[63088]: DEBUG oslo_concurrency.lockutils [None req-049e65c1-0c59-41b7-8448-5a3e6eb7d38f tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquiring lock "e845f60b-004f-4a5c-813e-0c183e99eba6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 827.906487] env[63088]: DEBUG oslo_concurrency.lockutils [None req-049e65c1-0c59-41b7-8448-5a3e6eb7d38f tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lock "e845f60b-004f-4a5c-813e-0c183e99eba6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 827.906651] env[63088]: DEBUG oslo_concurrency.lockutils [None req-049e65c1-0c59-41b7-8448-5a3e6eb7d38f tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lock "e845f60b-004f-4a5c-813e-0c183e99eba6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 827.908838] env[63088]: INFO nova.compute.manager [None req-049e65c1-0c59-41b7-8448-5a3e6eb7d38f tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: e845f60b-004f-4a5c-813e-0c183e99eba6] Terminating instance [ 827.910997] env[63088]: DEBUG nova.compute.manager [None req-049e65c1-0c59-41b7-8448-5a3e6eb7d38f tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: e845f60b-004f-4a5c-813e-0c183e99eba6] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 827.911170] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-049e65c1-0c59-41b7-8448-5a3e6eb7d38f tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: e845f60b-004f-4a5c-813e-0c183e99eba6] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 827.913024] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4234f092-3a6a-4fe1-899d-16bf086f6a2d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.920494] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-049e65c1-0c59-41b7-8448-5a3e6eb7d38f tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: e845f60b-004f-4a5c-813e-0c183e99eba6] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 827.920862] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6922fc98-32bf-467a-a8fa-d2ba4382831d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.927517] env[63088]: DEBUG oslo_vmware.api [None req-049e65c1-0c59-41b7-8448-5a3e6eb7d38f tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Waiting for the task: (returnval){ [ 827.927517] env[63088]: value = "task-1284902" [ 827.927517] env[63088]: _type = "Task" [ 827.927517] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.935581] env[63088]: DEBUG oslo_vmware.api [None req-049e65c1-0c59-41b7-8448-5a3e6eb7d38f tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1284902, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.012579] env[63088]: DEBUG oslo_vmware.api [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Task: {'id': task-1284897, 'name': PowerOnVM_Task, 'duration_secs': 0.534947} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.012940] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] [instance: 4f6e2d99-7300-42e6-8abf-362ddd4e8c6d] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 828.013304] env[63088]: INFO nova.compute.manager [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] [instance: 4f6e2d99-7300-42e6-8abf-362ddd4e8c6d] Took 8.57 seconds to spawn the instance on the hypervisor. [ 828.014238] env[63088]: DEBUG nova.compute.manager [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] [instance: 4f6e2d99-7300-42e6-8abf-362ddd4e8c6d] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 828.015201] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b3cc802-633e-4af4-9817-3cc9ee236083 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.033058] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 828.059956] env[63088]: DEBUG nova.compute.manager [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] [instance: 34ce411f-40c8-446e-b685-cd3ce07663dd] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 828.062158] env[63088]: DEBUG oslo_vmware.api [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Task: {'id': task-1284900, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.094839] env[63088]: DEBUG nova.virt.hardware [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 828.094839] env[63088]: DEBUG nova.virt.hardware [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 828.094839] env[63088]: DEBUG nova.virt.hardware [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 828.094993] env[63088]: DEBUG nova.virt.hardware [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 828.094993] env[63088]: DEBUG nova.virt.hardware [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 828.094993] env[63088]: DEBUG nova.virt.hardware [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 828.094993] env[63088]: DEBUG nova.virt.hardware [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 828.094993] env[63088]: DEBUG nova.virt.hardware [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 828.095144] env[63088]: DEBUG nova.virt.hardware [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 828.095144] env[63088]: DEBUG nova.virt.hardware [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 828.095144] env[63088]: DEBUG nova.virt.hardware [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 828.096476] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27f70eb6-3ef4-40cb-b30a-b60d10d44848 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.106298] env[63088]: DEBUG oslo_vmware.api [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Task: {'id': task-1284901, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.112539] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6744948-f15a-43a2-a197-9aa812c7b2b9 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.128712] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 828.366801] env[63088]: DEBUG oslo_concurrency.lockutils [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.340s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 828.367424] env[63088]: DEBUG nova.compute.manager [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] [instance: 1d421bf0-925d-408c-a612-b659da92799f] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 828.370297] env[63088]: DEBUG oslo_concurrency.lockutils [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.852s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 828.374047] env[63088]: INFO nova.compute.claims [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] [instance: 2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 828.440685] env[63088]: DEBUG oslo_vmware.api [None req-049e65c1-0c59-41b7-8448-5a3e6eb7d38f tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1284902, 'name': PowerOffVM_Task, 'duration_secs': 0.320605} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.441090] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-049e65c1-0c59-41b7-8448-5a3e6eb7d38f tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: e845f60b-004f-4a5c-813e-0c183e99eba6] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 828.441333] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-049e65c1-0c59-41b7-8448-5a3e6eb7d38f tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: e845f60b-004f-4a5c-813e-0c183e99eba6] Unregistering the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 828.441652] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8d6ac785-35f0-48db-8770-da1b9ed6ce00 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.526416] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-049e65c1-0c59-41b7-8448-5a3e6eb7d38f tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: e845f60b-004f-4a5c-813e-0c183e99eba6] Unregistered the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 828.526651] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-049e65c1-0c59-41b7-8448-5a3e6eb7d38f tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: e845f60b-004f-4a5c-813e-0c183e99eba6] Deleting contents of the VM from datastore datastore2 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 828.526888] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-049e65c1-0c59-41b7-8448-5a3e6eb7d38f tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Deleting the datastore file [datastore2] e845f60b-004f-4a5c-813e-0c183e99eba6 {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 828.531556] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5ae91e7a-51b5-4946-88cc-5723f76cb326 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.540248] env[63088]: DEBUG oslo_vmware.api [None req-049e65c1-0c59-41b7-8448-5a3e6eb7d38f tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Waiting for the task: (returnval){ [ 828.540248] env[63088]: value = "task-1284904" [ 828.540248] env[63088]: _type = "Task" [ 828.540248] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.540748] env[63088]: INFO nova.compute.manager [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] [instance: 4f6e2d99-7300-42e6-8abf-362ddd4e8c6d] Took 38.97 seconds to build instance. [ 828.550913] env[63088]: DEBUG oslo_vmware.api [None req-049e65c1-0c59-41b7-8448-5a3e6eb7d38f tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1284904, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.559317] env[63088]: DEBUG oslo_vmware.api [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Task: {'id': task-1284900, 'name': ReconfigVM_Task, 'duration_secs': 0.550002} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.559841] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] [instance: b159649b-ebd5-4c7d-9074-dc0e6395a947] Reconfigured VM instance instance-0000003b to attach disk [datastore1] b159649b-ebd5-4c7d-9074-dc0e6395a947/b159649b-ebd5-4c7d-9074-dc0e6395a947.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 828.560664] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-455919e9-01b1-4f05-a3af-835e39af2290 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.566682] env[63088]: DEBUG oslo_vmware.api [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Waiting for the task: (returnval){ [ 828.566682] env[63088]: value = "task-1284905" [ 828.566682] env[63088]: _type = "Task" [ 828.566682] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.574797] env[63088]: DEBUG oslo_vmware.api [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Task: {'id': task-1284905, 'name': Rename_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.595638] env[63088]: DEBUG oslo_vmware.api [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Task: {'id': task-1284901, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.732434} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.595944] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore2] 07fd3e6f-e0af-467c-9039-238be5c58d25/07fd3e6f-e0af-467c-9039-238be5c58d25.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 828.596190] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Extending root virtual disk to 1048576 {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 828.596443] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8b29494b-152a-4009-9a54-f37010bc9729 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.602504] env[63088]: DEBUG oslo_vmware.api [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Waiting for the task: (returnval){ [ 828.602504] env[63088]: value = "task-1284906" [ 828.602504] env[63088]: _type = "Task" [ 828.602504] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.609832] env[63088]: DEBUG oslo_vmware.api [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Task: {'id': task-1284906, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.672887] env[63088]: DEBUG nova.compute.manager [req-efef5655-7d76-4af3-912d-d73bf5446d27 req-df9e08a2-730d-4550-b9e0-fa60ae63443a service nova] [instance: 34ce411f-40c8-446e-b685-cd3ce07663dd] Received event network-vif-plugged-44d58f1e-f90f-4aea-b8fd-5f9c7ff195cb {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 828.673147] env[63088]: DEBUG oslo_concurrency.lockutils [req-efef5655-7d76-4af3-912d-d73bf5446d27 req-df9e08a2-730d-4550-b9e0-fa60ae63443a service nova] Acquiring lock "34ce411f-40c8-446e-b685-cd3ce07663dd-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 828.673353] env[63088]: DEBUG oslo_concurrency.lockutils [req-efef5655-7d76-4af3-912d-d73bf5446d27 req-df9e08a2-730d-4550-b9e0-fa60ae63443a service nova] Lock "34ce411f-40c8-446e-b685-cd3ce07663dd-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 828.673547] env[63088]: DEBUG oslo_concurrency.lockutils [req-efef5655-7d76-4af3-912d-d73bf5446d27 req-df9e08a2-730d-4550-b9e0-fa60ae63443a service nova] Lock "34ce411f-40c8-446e-b685-cd3ce07663dd-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 828.673754] env[63088]: DEBUG nova.compute.manager [req-efef5655-7d76-4af3-912d-d73bf5446d27 req-df9e08a2-730d-4550-b9e0-fa60ae63443a service nova] [instance: 34ce411f-40c8-446e-b685-cd3ce07663dd] No waiting events found dispatching network-vif-plugged-44d58f1e-f90f-4aea-b8fd-5f9c7ff195cb {{(pid=63088) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 828.673992] env[63088]: WARNING nova.compute.manager [req-efef5655-7d76-4af3-912d-d73bf5446d27 req-df9e08a2-730d-4550-b9e0-fa60ae63443a service nova] [instance: 34ce411f-40c8-446e-b685-cd3ce07663dd] Received unexpected event network-vif-plugged-44d58f1e-f90f-4aea-b8fd-5f9c7ff195cb for instance with vm_state building and task_state spawning. [ 828.710585] env[63088]: DEBUG nova.virt.hardware [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 828.710885] env[63088]: DEBUG nova.virt.hardware [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 828.711213] env[63088]: DEBUG nova.virt.hardware [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 828.711457] env[63088]: DEBUG nova.virt.hardware [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 828.711612] env[63088]: DEBUG nova.virt.hardware [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 828.711763] env[63088]: DEBUG nova.virt.hardware [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 828.711975] env[63088]: DEBUG nova.virt.hardware [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 828.712152] env[63088]: DEBUG nova.virt.hardware [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 828.712323] env[63088]: DEBUG nova.virt.hardware [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 828.712492] env[63088]: DEBUG nova.virt.hardware [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 828.712665] env[63088]: DEBUG nova.virt.hardware [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 828.713878] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-260b58d1-1968-4b44-bf41-f10f73f01242 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.722536] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31165070-6f5b-40b2-a9a7-a58c3abbbda0 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.738178] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6d:a9:02', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '21310d90-efbc-45a8-a97f-c4358606530f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '53b4bd8d-df8b-4069-a75a-af46b605dddb', 'vif_model': 'vmxnet3'}] {{(pid=63088) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 828.747241] env[63088]: DEBUG oslo.service.loopingcall [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 828.747813] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] Creating VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 828.748171] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-842a8b53-b685-40c1-a84c-5083c75ddcde {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.768538] env[63088]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 828.768538] env[63088]: value = "task-1284907" [ 828.768538] env[63088]: _type = "Task" [ 828.768538] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.775991] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1284907, 'name': CreateVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.875804] env[63088]: DEBUG nova.compute.utils [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 828.881265] env[63088]: DEBUG nova.compute.manager [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] [instance: 1d421bf0-925d-408c-a612-b659da92799f] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 828.881483] env[63088]: DEBUG nova.network.neutron [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] [instance: 1d421bf0-925d-408c-a612-b659da92799f] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 828.911106] env[63088]: DEBUG nova.network.neutron [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] [instance: 34ce411f-40c8-446e-b685-cd3ce07663dd] Successfully updated port: 44d58f1e-f90f-4aea-b8fd-5f9c7ff195cb {{(pid=63088) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 828.962677] env[63088]: DEBUG nova.policy [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ca8c6f64e1724943aaef6e07fa3b7308', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'edf7d041b681480791590601dde8ba97', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 829.047372] env[63088]: DEBUG oslo_concurrency.lockutils [None req-81eabb7b-b27e-4918-b643-91256d7cb598 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Lock "4f6e2d99-7300-42e6-8abf-362ddd4e8c6d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 74.318s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 829.055158] env[63088]: DEBUG oslo_vmware.api [None req-049e65c1-0c59-41b7-8448-5a3e6eb7d38f tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1284904, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.25627} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.055406] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-049e65c1-0c59-41b7-8448-5a3e6eb7d38f tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 829.055595] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-049e65c1-0c59-41b7-8448-5a3e6eb7d38f tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: e845f60b-004f-4a5c-813e-0c183e99eba6] Deleted contents of the VM from datastore datastore2 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 829.055775] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-049e65c1-0c59-41b7-8448-5a3e6eb7d38f tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: e845f60b-004f-4a5c-813e-0c183e99eba6] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 829.055943] env[63088]: INFO nova.compute.manager [None req-049e65c1-0c59-41b7-8448-5a3e6eb7d38f tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: e845f60b-004f-4a5c-813e-0c183e99eba6] Took 1.14 seconds to destroy the instance on the hypervisor. [ 829.056209] env[63088]: DEBUG oslo.service.loopingcall [None req-049e65c1-0c59-41b7-8448-5a3e6eb7d38f tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 829.056399] env[63088]: DEBUG nova.compute.manager [-] [instance: e845f60b-004f-4a5c-813e-0c183e99eba6] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 829.056495] env[63088]: DEBUG nova.network.neutron [-] [instance: e845f60b-004f-4a5c-813e-0c183e99eba6] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 829.076808] env[63088]: DEBUG oslo_vmware.api [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Task: {'id': task-1284905, 'name': Rename_Task, 'duration_secs': 0.136082} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.076808] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] [instance: b159649b-ebd5-4c7d-9074-dc0e6395a947] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 829.076985] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0462ed8e-4bd4-4d4f-8112-085a1abcb7e7 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.083271] env[63088]: DEBUG oslo_vmware.api [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Waiting for the task: (returnval){ [ 829.083271] env[63088]: value = "task-1284908" [ 829.083271] env[63088]: _type = "Task" [ 829.083271] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.091928] env[63088]: DEBUG oslo_vmware.api [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Task: {'id': task-1284908, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.112337] env[63088]: DEBUG oslo_vmware.api [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Task: {'id': task-1284906, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065521} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.112601] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Extended root virtual disk {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 829.113404] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eac0c001-8ca5-432b-925c-b58f68014603 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.135899] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Reconfiguring VM instance instance-0000003a to attach disk [datastore2] 07fd3e6f-e0af-467c-9039-238be5c58d25/07fd3e6f-e0af-467c-9039-238be5c58d25.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 829.136778] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-39f06010-5553-4e4b-8ebb-4ba1a7705418 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.159275] env[63088]: DEBUG oslo_vmware.api [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Waiting for the task: (returnval){ [ 829.159275] env[63088]: value = "task-1284909" [ 829.159275] env[63088]: _type = "Task" [ 829.159275] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.167627] env[63088]: DEBUG oslo_vmware.api [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Task: {'id': task-1284909, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.278929] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1284907, 'name': CreateVM_Task} progress is 99%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.389754] env[63088]: DEBUG nova.compute.manager [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] [instance: 1d421bf0-925d-408c-a612-b659da92799f] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 829.414712] env[63088]: DEBUG oslo_concurrency.lockutils [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Acquiring lock "refresh_cache-34ce411f-40c8-446e-b685-cd3ce07663dd" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 829.414875] env[63088]: DEBUG oslo_concurrency.lockutils [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Acquired lock "refresh_cache-34ce411f-40c8-446e-b685-cd3ce07663dd" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 829.415107] env[63088]: DEBUG nova.network.neutron [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] [instance: 34ce411f-40c8-446e-b685-cd3ce07663dd] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 829.437052] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a26b18b9-d003-4a4b-9d6c-96d60267d634 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Acquiring lock "4f6e2d99-7300-42e6-8abf-362ddd4e8c6d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 829.437328] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a26b18b9-d003-4a4b-9d6c-96d60267d634 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Lock "4f6e2d99-7300-42e6-8abf-362ddd4e8c6d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.441020] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a26b18b9-d003-4a4b-9d6c-96d60267d634 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Acquiring lock "4f6e2d99-7300-42e6-8abf-362ddd4e8c6d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 829.441020] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a26b18b9-d003-4a4b-9d6c-96d60267d634 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Lock "4f6e2d99-7300-42e6-8abf-362ddd4e8c6d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.441020] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a26b18b9-d003-4a4b-9d6c-96d60267d634 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Lock "4f6e2d99-7300-42e6-8abf-362ddd4e8c6d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 829.443275] env[63088]: INFO nova.compute.manager [None req-a26b18b9-d003-4a4b-9d6c-96d60267d634 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] [instance: 4f6e2d99-7300-42e6-8abf-362ddd4e8c6d] Terminating instance [ 829.448775] env[63088]: DEBUG nova.compute.manager [None req-a26b18b9-d003-4a4b-9d6c-96d60267d634 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] [instance: 4f6e2d99-7300-42e6-8abf-362ddd4e8c6d] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 829.448975] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-a26b18b9-d003-4a4b-9d6c-96d60267d634 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] [instance: 4f6e2d99-7300-42e6-8abf-362ddd4e8c6d] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 829.449973] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4325031c-9ef3-40c0-a616-1944e0e4cc9f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.458403] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-a26b18b9-d003-4a4b-9d6c-96d60267d634 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] [instance: 4f6e2d99-7300-42e6-8abf-362ddd4e8c6d] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 829.461025] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c6855a04-3dc8-48df-8e4d-9d01eb41cc17 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.468325] env[63088]: DEBUG oslo_vmware.api [None req-a26b18b9-d003-4a4b-9d6c-96d60267d634 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Waiting for the task: (returnval){ [ 829.468325] env[63088]: value = "task-1284910" [ 829.468325] env[63088]: _type = "Task" [ 829.468325] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.478578] env[63088]: DEBUG oslo_vmware.api [None req-a26b18b9-d003-4a4b-9d6c-96d60267d634 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Task: {'id': task-1284910, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.552138] env[63088]: DEBUG nova.compute.manager [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 829.597553] env[63088]: DEBUG oslo_vmware.api [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Task: {'id': task-1284908, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.670986] env[63088]: DEBUG oslo_vmware.api [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Task: {'id': task-1284909, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.699032] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8704d360-2fff-4754-b976-c4bbe91c8c92 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.711283] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e311d21-1c95-4b23-968c-e40fd7acd8f2 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.743758] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b253dd61-b744-4095-8871-6a58aed69f6d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.751686] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c654595c-89da-48d4-b11e-9a947b22a6b8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.766609] env[63088]: DEBUG nova.compute.provider_tree [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 829.778653] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1284907, 'name': CreateVM_Task, 'duration_secs': 0.561298} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.778835] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] Created VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 829.779652] env[63088]: DEBUG oslo_concurrency.lockutils [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 829.779766] env[63088]: DEBUG oslo_concurrency.lockutils [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 829.780059] env[63088]: DEBUG oslo_concurrency.lockutils [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 829.780323] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f4e4a91f-c2f2-44dc-b70c-fc76125b8175 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.785406] env[63088]: DEBUG oslo_vmware.api [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Waiting for the task: (returnval){ [ 829.785406] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52de6666-5557-367b-7e09-896d628319be" [ 829.785406] env[63088]: _type = "Task" [ 829.785406] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.794329] env[63088]: DEBUG oslo_vmware.api [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52de6666-5557-367b-7e09-896d628319be, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.968420] env[63088]: DEBUG nova.network.neutron [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] [instance: 1d421bf0-925d-408c-a612-b659da92799f] Successfully created port: 1c83da13-18b2-45c9-bc41-4d45e17606d5 {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 829.980009] env[63088]: DEBUG oslo_vmware.api [None req-a26b18b9-d003-4a4b-9d6c-96d60267d634 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Task: {'id': task-1284910, 'name': PowerOffVM_Task, 'duration_secs': 0.181412} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.980284] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-a26b18b9-d003-4a4b-9d6c-96d60267d634 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] [instance: 4f6e2d99-7300-42e6-8abf-362ddd4e8c6d] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 829.980459] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-a26b18b9-d003-4a4b-9d6c-96d60267d634 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] [instance: 4f6e2d99-7300-42e6-8abf-362ddd4e8c6d] Unregistering the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 829.980758] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8541c02b-cf41-4e28-bd2b-72d16e2290d7 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.988545] env[63088]: DEBUG nova.network.neutron [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] [instance: 34ce411f-40c8-446e-b685-cd3ce07663dd] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 830.038548] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-a26b18b9-d003-4a4b-9d6c-96d60267d634 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] [instance: 4f6e2d99-7300-42e6-8abf-362ddd4e8c6d] Unregistered the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 830.038958] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-a26b18b9-d003-4a4b-9d6c-96d60267d634 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] [instance: 4f6e2d99-7300-42e6-8abf-362ddd4e8c6d] Deleting contents of the VM from datastore datastore2 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 830.039264] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-a26b18b9-d003-4a4b-9d6c-96d60267d634 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Deleting the datastore file [datastore2] 4f6e2d99-7300-42e6-8abf-362ddd4e8c6d {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 830.039662] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-522e5e8b-2e01-40fe-9245-8a9e671d5cda {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.050203] env[63088]: DEBUG oslo_vmware.api [None req-a26b18b9-d003-4a4b-9d6c-96d60267d634 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Waiting for the task: (returnval){ [ 830.050203] env[63088]: value = "task-1284912" [ 830.050203] env[63088]: _type = "Task" [ 830.050203] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.064808] env[63088]: DEBUG oslo_vmware.api [None req-a26b18b9-d003-4a4b-9d6c-96d60267d634 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Task: {'id': task-1284912, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.083418] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 830.093245] env[63088]: DEBUG oslo_vmware.api [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Task: {'id': task-1284908, 'name': PowerOnVM_Task, 'duration_secs': 0.686329} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.096726] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] [instance: b159649b-ebd5-4c7d-9074-dc0e6395a947] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 830.096976] env[63088]: INFO nova.compute.manager [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] [instance: b159649b-ebd5-4c7d-9074-dc0e6395a947] Took 6.35 seconds to spawn the instance on the hypervisor. [ 830.097189] env[63088]: DEBUG nova.compute.manager [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] [instance: b159649b-ebd5-4c7d-9074-dc0e6395a947] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 830.098103] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12b222cd-9637-4d3a-a96d-56e5c707da59 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.172481] env[63088]: DEBUG oslo_vmware.api [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Task: {'id': task-1284909, 'name': ReconfigVM_Task, 'duration_secs': 0.573368} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.175486] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Reconfigured VM instance instance-0000003a to attach disk [datastore2] 07fd3e6f-e0af-467c-9039-238be5c58d25/07fd3e6f-e0af-467c-9039-238be5c58d25.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 830.176564] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ece16035-b05b-4ced-aac4-74381f129441 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.184505] env[63088]: DEBUG oslo_vmware.api [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Waiting for the task: (returnval){ [ 830.184505] env[63088]: value = "task-1284913" [ 830.184505] env[63088]: _type = "Task" [ 830.184505] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.194382] env[63088]: DEBUG oslo_vmware.api [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Task: {'id': task-1284913, 'name': Rename_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.270069] env[63088]: DEBUG nova.scheduler.client.report [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 830.297487] env[63088]: DEBUG oslo_vmware.api [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52de6666-5557-367b-7e09-896d628319be, 'name': SearchDatastore_Task, 'duration_secs': 0.009172} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.297794] env[63088]: DEBUG oslo_concurrency.lockutils [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 830.298053] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] Processing image 1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 830.298291] env[63088]: DEBUG oslo_concurrency.lockutils [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.298471] env[63088]: DEBUG oslo_concurrency.lockutils [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.298706] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 830.298978] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-be0cc842-48d0-4e51-9366-1d85f2c1386f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.308469] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 830.308659] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 830.309505] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-07ac4108-430f-4fa6-ae63-f47ec8033b0b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.315117] env[63088]: DEBUG oslo_vmware.api [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Waiting for the task: (returnval){ [ 830.315117] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52086906-e1bb-00a4-d172-34609f6f602e" [ 830.315117] env[63088]: _type = "Task" [ 830.315117] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.323636] env[63088]: DEBUG oslo_vmware.api [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52086906-e1bb-00a4-d172-34609f6f602e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.380386] env[63088]: DEBUG nova.network.neutron [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] [instance: 34ce411f-40c8-446e-b685-cd3ce07663dd] Updating instance_info_cache with network_info: [{"id": "44d58f1e-f90f-4aea-b8fd-5f9c7ff195cb", "address": "fa:16:3e:bb:b4:13", "network": {"id": "88a981da-0420-42f7-bc5c-0bfb53e39272", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-1197986229-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a5f62816dd0243da9be2073086b136d8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92fe29b3-0907-453d-aabb-5559c4bd7c0f", "external-id": "nsx-vlan-transportzone-482", "segmentation_id": 482, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap44d58f1e-f9", "ovs_interfaceid": "44d58f1e-f90f-4aea-b8fd-5f9c7ff195cb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 830.403418] env[63088]: DEBUG nova.compute.manager [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] [instance: 1d421bf0-925d-408c-a612-b659da92799f] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 830.407533] env[63088]: DEBUG nova.network.neutron [-] [instance: e845f60b-004f-4a5c-813e-0c183e99eba6] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 830.434808] env[63088]: DEBUG nova.virt.hardware [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 830.435052] env[63088]: DEBUG nova.virt.hardware [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 830.435215] env[63088]: DEBUG nova.virt.hardware [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 830.435394] env[63088]: DEBUG nova.virt.hardware [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 830.435541] env[63088]: DEBUG nova.virt.hardware [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 830.435690] env[63088]: DEBUG nova.virt.hardware [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 830.435898] env[63088]: DEBUG nova.virt.hardware [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 830.436246] env[63088]: DEBUG nova.virt.hardware [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 830.436535] env[63088]: DEBUG nova.virt.hardware [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 830.436764] env[63088]: DEBUG nova.virt.hardware [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 830.436794] env[63088]: DEBUG nova.virt.hardware [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 830.438288] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74e39c85-b1cd-42f5-8fa2-245cc323111c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.447588] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41985169-3f66-4fcb-a236-25b69e1b4511 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.561974] env[63088]: DEBUG oslo_vmware.api [None req-a26b18b9-d003-4a4b-9d6c-96d60267d634 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Task: {'id': task-1284912, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.191087} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.561974] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-a26b18b9-d003-4a4b-9d6c-96d60267d634 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 830.561974] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-a26b18b9-d003-4a4b-9d6c-96d60267d634 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] [instance: 4f6e2d99-7300-42e6-8abf-362ddd4e8c6d] Deleted contents of the VM from datastore datastore2 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 830.561974] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-a26b18b9-d003-4a4b-9d6c-96d60267d634 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] [instance: 4f6e2d99-7300-42e6-8abf-362ddd4e8c6d] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 830.562230] env[63088]: INFO nova.compute.manager [None req-a26b18b9-d003-4a4b-9d6c-96d60267d634 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] [instance: 4f6e2d99-7300-42e6-8abf-362ddd4e8c6d] Took 1.11 seconds to destroy the instance on the hypervisor. [ 830.562308] env[63088]: DEBUG oslo.service.loopingcall [None req-a26b18b9-d003-4a4b-9d6c-96d60267d634 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 830.562505] env[63088]: DEBUG nova.compute.manager [-] [instance: 4f6e2d99-7300-42e6-8abf-362ddd4e8c6d] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 830.562609] env[63088]: DEBUG nova.network.neutron [-] [instance: 4f6e2d99-7300-42e6-8abf-362ddd4e8c6d] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 830.617978] env[63088]: INFO nova.compute.manager [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] [instance: b159649b-ebd5-4c7d-9074-dc0e6395a947] Took 34.99 seconds to build instance. [ 830.696604] env[63088]: DEBUG oslo_vmware.api [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Task: {'id': task-1284913, 'name': Rename_Task, 'duration_secs': 0.145422} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.697169] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 830.697364] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a4412c7b-13f7-4aa6-8099-e183730cc2ce {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.703571] env[63088]: DEBUG oslo_vmware.api [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Waiting for the task: (returnval){ [ 830.703571] env[63088]: value = "task-1284914" [ 830.703571] env[63088]: _type = "Task" [ 830.703571] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.714384] env[63088]: DEBUG oslo_vmware.api [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Task: {'id': task-1284914, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.737847] env[63088]: DEBUG nova.compute.manager [req-c9634819-d34c-4d13-b193-0287b79180c6 req-eb18b367-490d-4003-9ef8-eeac955cbc4d service nova] [instance: 34ce411f-40c8-446e-b685-cd3ce07663dd] Received event network-changed-44d58f1e-f90f-4aea-b8fd-5f9c7ff195cb {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 830.738219] env[63088]: DEBUG nova.compute.manager [req-c9634819-d34c-4d13-b193-0287b79180c6 req-eb18b367-490d-4003-9ef8-eeac955cbc4d service nova] [instance: 34ce411f-40c8-446e-b685-cd3ce07663dd] Refreshing instance network info cache due to event network-changed-44d58f1e-f90f-4aea-b8fd-5f9c7ff195cb. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 830.738603] env[63088]: DEBUG oslo_concurrency.lockutils [req-c9634819-d34c-4d13-b193-0287b79180c6 req-eb18b367-490d-4003-9ef8-eeac955cbc4d service nova] Acquiring lock "refresh_cache-34ce411f-40c8-446e-b685-cd3ce07663dd" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.779202] env[63088]: DEBUG oslo_concurrency.lockutils [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.408s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 830.779202] env[63088]: DEBUG nova.compute.manager [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] [instance: 2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 830.788939] env[63088]: DEBUG oslo_concurrency.lockutils [None req-da2e38c8-9951-4dcc-950e-b6d060a5ab7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.594s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 830.788939] env[63088]: DEBUG nova.objects.instance [None req-da2e38c8-9951-4dcc-950e-b6d060a5ab7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lazy-loading 'resources' on Instance uuid 11ec987f-9908-4366-81ae-cbfe4de67b2d {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 830.825582] env[63088]: DEBUG oslo_vmware.api [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52086906-e1bb-00a4-d172-34609f6f602e, 'name': SearchDatastore_Task, 'duration_secs': 0.009155} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.826431] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-092f2897-2dab-4643-872d-6c1d2993860b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.833805] env[63088]: DEBUG oslo_vmware.api [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Waiting for the task: (returnval){ [ 830.833805] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52c15c5e-74f9-79c4-ede6-d92ac0b918cc" [ 830.833805] env[63088]: _type = "Task" [ 830.833805] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.843047] env[63088]: DEBUG oslo_vmware.api [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52c15c5e-74f9-79c4-ede6-d92ac0b918cc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.884703] env[63088]: DEBUG oslo_concurrency.lockutils [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Releasing lock "refresh_cache-34ce411f-40c8-446e-b685-cd3ce07663dd" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 830.885089] env[63088]: DEBUG nova.compute.manager [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] [instance: 34ce411f-40c8-446e-b685-cd3ce07663dd] Instance network_info: |[{"id": "44d58f1e-f90f-4aea-b8fd-5f9c7ff195cb", "address": "fa:16:3e:bb:b4:13", "network": {"id": "88a981da-0420-42f7-bc5c-0bfb53e39272", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-1197986229-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a5f62816dd0243da9be2073086b136d8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92fe29b3-0907-453d-aabb-5559c4bd7c0f", "external-id": "nsx-vlan-transportzone-482", "segmentation_id": 482, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap44d58f1e-f9", "ovs_interfaceid": "44d58f1e-f90f-4aea-b8fd-5f9c7ff195cb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 830.885450] env[63088]: DEBUG oslo_concurrency.lockutils [req-c9634819-d34c-4d13-b193-0287b79180c6 req-eb18b367-490d-4003-9ef8-eeac955cbc4d service nova] Acquired lock "refresh_cache-34ce411f-40c8-446e-b685-cd3ce07663dd" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.885673] env[63088]: DEBUG nova.network.neutron [req-c9634819-d34c-4d13-b193-0287b79180c6 req-eb18b367-490d-4003-9ef8-eeac955cbc4d service nova] [instance: 34ce411f-40c8-446e-b685-cd3ce07663dd] Refreshing network info cache for port 44d58f1e-f90f-4aea-b8fd-5f9c7ff195cb {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 830.886875] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] [instance: 34ce411f-40c8-446e-b685-cd3ce07663dd] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:bb:b4:13', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '92fe29b3-0907-453d-aabb-5559c4bd7c0f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '44d58f1e-f90f-4aea-b8fd-5f9c7ff195cb', 'vif_model': 'vmxnet3'}] {{(pid=63088) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 830.898290] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Creating folder: Project (a5f62816dd0243da9be2073086b136d8). Parent ref: group-v275816. {{(pid=63088) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 830.903646] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-205be3b5-3210-4ff7-954d-d06a00cdbcbd {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.908456] env[63088]: INFO nova.compute.manager [-] [instance: e845f60b-004f-4a5c-813e-0c183e99eba6] Took 1.85 seconds to deallocate network for instance. [ 830.917043] env[63088]: INFO nova.virt.vmwareapi.vm_util [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Created folder: Project (a5f62816dd0243da9be2073086b136d8) in parent group-v275816. [ 830.917043] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Creating folder: Instances. Parent ref: group-v275858. {{(pid=63088) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 830.917043] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7a35c590-ab0f-460d-b5d6-11feb4ba4ddd {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.926761] env[63088]: INFO nova.virt.vmwareapi.vm_util [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Created folder: Instances in parent group-v275858. [ 830.927083] env[63088]: DEBUG oslo.service.loopingcall [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 830.927674] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 34ce411f-40c8-446e-b685-cd3ce07663dd] Creating VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 830.927674] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-260677e9-be60-4d0e-b609-5ad1b1ab881a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.951644] env[63088]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 830.951644] env[63088]: value = "task-1284917" [ 830.951644] env[63088]: _type = "Task" [ 830.951644] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.959833] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1284917, 'name': CreateVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.120662] env[63088]: DEBUG oslo_concurrency.lockutils [None req-72bfc1ca-96b4-4c1e-bb59-e212d903b167 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Lock "b159649b-ebd5-4c7d-9074-dc0e6395a947" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 67.674s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 831.216457] env[63088]: DEBUG oslo_vmware.api [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Task: {'id': task-1284914, 'name': PowerOnVM_Task, 'duration_secs': 0.499481} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.216790] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 831.217071] env[63088]: INFO nova.compute.manager [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Took 9.86 seconds to spawn the instance on the hypervisor. [ 831.217965] env[63088]: DEBUG nova.compute.manager [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 831.221759] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e6102c4-184b-4e77-9bd1-24c7329ed84c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.292059] env[63088]: DEBUG nova.compute.utils [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 831.292059] env[63088]: DEBUG nova.compute.manager [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] [instance: 2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 831.292059] env[63088]: DEBUG nova.network.neutron [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] [instance: 2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 831.313068] env[63088]: DEBUG nova.network.neutron [req-c9634819-d34c-4d13-b193-0287b79180c6 req-eb18b367-490d-4003-9ef8-eeac955cbc4d service nova] [instance: 34ce411f-40c8-446e-b685-cd3ce07663dd] Updated VIF entry in instance network info cache for port 44d58f1e-f90f-4aea-b8fd-5f9c7ff195cb. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 831.313497] env[63088]: DEBUG nova.network.neutron [req-c9634819-d34c-4d13-b193-0287b79180c6 req-eb18b367-490d-4003-9ef8-eeac955cbc4d service nova] [instance: 34ce411f-40c8-446e-b685-cd3ce07663dd] Updating instance_info_cache with network_info: [{"id": "44d58f1e-f90f-4aea-b8fd-5f9c7ff195cb", "address": "fa:16:3e:bb:b4:13", "network": {"id": "88a981da-0420-42f7-bc5c-0bfb53e39272", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-1197986229-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a5f62816dd0243da9be2073086b136d8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92fe29b3-0907-453d-aabb-5559c4bd7c0f", "external-id": "nsx-vlan-transportzone-482", "segmentation_id": 482, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap44d58f1e-f9", "ovs_interfaceid": "44d58f1e-f90f-4aea-b8fd-5f9c7ff195cb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 831.343317] env[63088]: DEBUG nova.policy [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '99446d9d9bf140f8b2c248713131e1af', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6bd87f64aa654aad8f632c4f22a2d266', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 831.351233] env[63088]: DEBUG oslo_vmware.api [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52c15c5e-74f9-79c4-ede6-d92ac0b918cc, 'name': SearchDatastore_Task, 'duration_secs': 0.010755} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.351503] env[63088]: DEBUG oslo_concurrency.lockutils [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 831.351808] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] 4847f15e-98d4-401c-91ff-067e84a85727/4847f15e-98d4-401c-91ff-067e84a85727.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 831.352091] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cb76f104-1647-4fda-904e-f24809df5b05 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.359902] env[63088]: DEBUG oslo_vmware.api [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Waiting for the task: (returnval){ [ 831.359902] env[63088]: value = "task-1284918" [ 831.359902] env[63088]: _type = "Task" [ 831.359902] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.367364] env[63088]: DEBUG oslo_vmware.api [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1284918, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.387822] env[63088]: INFO nova.compute.manager [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] [instance: b159649b-ebd5-4c7d-9074-dc0e6395a947] Rebuilding instance [ 831.415229] env[63088]: DEBUG oslo_concurrency.lockutils [None req-049e65c1-0c59-41b7-8448-5a3e6eb7d38f tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.418169] env[63088]: DEBUG nova.network.neutron [-] [instance: 4f6e2d99-7300-42e6-8abf-362ddd4e8c6d] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 831.443108] env[63088]: DEBUG nova.compute.manager [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] [instance: b159649b-ebd5-4c7d-9074-dc0e6395a947] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 831.444102] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89179b08-3e24-44ba-a9d5-f7a8b1ae5ae9 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.464595] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1284917, 'name': CreateVM_Task, 'duration_secs': 0.363503} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.467194] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 34ce411f-40c8-446e-b685-cd3ce07663dd] Created VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 831.468263] env[63088]: DEBUG oslo_concurrency.lockutils [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 831.468501] env[63088]: DEBUG oslo_concurrency.lockutils [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.468838] env[63088]: DEBUG oslo_concurrency.lockutils [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 831.469411] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-851abc72-88ae-419a-b80c-ddc3784876ea {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.474444] env[63088]: DEBUG oslo_vmware.api [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Waiting for the task: (returnval){ [ 831.474444] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52cdf995-649d-ec00-ae6a-a921e098c123" [ 831.474444] env[63088]: _type = "Task" [ 831.474444] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.483672] env[63088]: DEBUG oslo_vmware.api [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52cdf995-649d-ec00-ae6a-a921e098c123, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.570200] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49440104-2798-4cbe-8c95-7c65aa5666e9 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.578218] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f8e2bca-c8d4-46fa-b049-8dc34a99d45e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.610274] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2456cc15-d939-44a7-94e1-008a1d7843e1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.619432] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4abeba8b-680f-45b3-865b-16cfa3e75ffe {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.624622] env[63088]: DEBUG nova.compute.manager [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 831.637527] env[63088]: DEBUG nova.compute.provider_tree [None req-da2e38c8-9951-4dcc-950e-b6d060a5ab7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 831.654838] env[63088]: DEBUG nova.network.neutron [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] [instance: 2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68] Successfully created port: 4c03d05a-bb4d-4d2b-bb2c-65f92a25278c {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 831.738591] env[63088]: INFO nova.compute.manager [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Took 40.28 seconds to build instance. [ 831.796390] env[63088]: DEBUG nova.compute.manager [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] [instance: 2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 831.817091] env[63088]: DEBUG oslo_concurrency.lockutils [req-c9634819-d34c-4d13-b193-0287b79180c6 req-eb18b367-490d-4003-9ef8-eeac955cbc4d service nova] Releasing lock "refresh_cache-34ce411f-40c8-446e-b685-cd3ce07663dd" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 831.817365] env[63088]: DEBUG nova.compute.manager [req-c9634819-d34c-4d13-b193-0287b79180c6 req-eb18b367-490d-4003-9ef8-eeac955cbc4d service nova] [instance: e845f60b-004f-4a5c-813e-0c183e99eba6] Received event network-vif-deleted-8ca1e82d-550d-4faa-8b7e-93e0b812c234 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 831.871213] env[63088]: DEBUG oslo_vmware.api [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1284918, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.922236] env[63088]: INFO nova.compute.manager [-] [instance: 4f6e2d99-7300-42e6-8abf-362ddd4e8c6d] Took 1.36 seconds to deallocate network for instance. [ 831.960929] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] [instance: b159649b-ebd5-4c7d-9074-dc0e6395a947] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 831.961664] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fdd400b1-b00d-4f74-9814-d31a6442eaee {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.970171] env[63088]: DEBUG oslo_vmware.api [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Waiting for the task: (returnval){ [ 831.970171] env[63088]: value = "task-1284919" [ 831.970171] env[63088]: _type = "Task" [ 831.970171] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.985873] env[63088]: DEBUG oslo_vmware.api [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Task: {'id': task-1284919, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.991189] env[63088]: DEBUG oslo_vmware.api [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52cdf995-649d-ec00-ae6a-a921e098c123, 'name': SearchDatastore_Task, 'duration_secs': 0.021659} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.991594] env[63088]: DEBUG oslo_concurrency.lockutils [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 831.991853] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] [instance: 34ce411f-40c8-446e-b685-cd3ce07663dd] Processing image 1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 831.992223] env[63088]: DEBUG oslo_concurrency.lockutils [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 831.992420] env[63088]: DEBUG oslo_concurrency.lockutils [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.992566] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 831.996608] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b4b9e57b-3f68-43a6-9ac6-b93af608323b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.011088] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 832.011346] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 832.012315] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-550a5ab7-a315-41cb-8f82-bbb33932dc26 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.022652] env[63088]: DEBUG oslo_vmware.api [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Waiting for the task: (returnval){ [ 832.022652] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52d46c04-8d14-0d5c-c6c6-af92a57c5650" [ 832.022652] env[63088]: _type = "Task" [ 832.022652] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.034014] env[63088]: DEBUG oslo_vmware.api [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52d46c04-8d14-0d5c-c6c6-af92a57c5650, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.140671] env[63088]: DEBUG nova.scheduler.client.report [None req-da2e38c8-9951-4dcc-950e-b6d060a5ab7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 832.157923] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 832.236293] env[63088]: DEBUG nova.network.neutron [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] [instance: 1d421bf0-925d-408c-a612-b659da92799f] Successfully updated port: 1c83da13-18b2-45c9-bc41-4d45e17606d5 {{(pid=63088) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 832.242055] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bd29667f-3ff2-4db8-8e61-dc7553b5924d tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Lock "07fd3e6f-e0af-467c-9039-238be5c58d25" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 76.734s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 832.372858] env[63088]: DEBUG oslo_vmware.api [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1284918, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.896104} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.373173] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] 4847f15e-98d4-401c-91ff-067e84a85727/4847f15e-98d4-401c-91ff-067e84a85727.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 832.373402] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] Extending root virtual disk to 1048576 {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 832.373712] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2f3a340d-6944-4a4e-9248-e26ebefe0854 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.380627] env[63088]: DEBUG oslo_vmware.api [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Waiting for the task: (returnval){ [ 832.380627] env[63088]: value = "task-1284920" [ 832.380627] env[63088]: _type = "Task" [ 832.380627] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.388470] env[63088]: DEBUG oslo_vmware.api [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1284920, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.430302] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a26b18b9-d003-4a4b-9d6c-96d60267d634 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 832.481487] env[63088]: DEBUG oslo_vmware.api [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Task: {'id': task-1284919, 'name': PowerOffVM_Task, 'duration_secs': 0.132101} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.482249] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] [instance: b159649b-ebd5-4c7d-9074-dc0e6395a947] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 832.482472] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] [instance: b159649b-ebd5-4c7d-9074-dc0e6395a947] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 832.483286] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f79bd1e3-e543-4e71-8b01-8c4d8a0c6264 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.493418] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] [instance: b159649b-ebd5-4c7d-9074-dc0e6395a947] Unregistering the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 832.493418] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1cfb81b9-53e0-4108-b5d8-300a48a192a7 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.517411] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] [instance: b159649b-ebd5-4c7d-9074-dc0e6395a947] Unregistered the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 832.517702] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] [instance: b159649b-ebd5-4c7d-9074-dc0e6395a947] Deleting contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 832.517924] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Deleting the datastore file [datastore1] b159649b-ebd5-4c7d-9074-dc0e6395a947 {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 832.518441] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3ebffe54-6544-4e73-bba2-c46d6ae2054e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.527223] env[63088]: DEBUG oslo_vmware.api [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Waiting for the task: (returnval){ [ 832.527223] env[63088]: value = "task-1284922" [ 832.527223] env[63088]: _type = "Task" [ 832.527223] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.533910] env[63088]: DEBUG oslo_vmware.api [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52d46c04-8d14-0d5c-c6c6-af92a57c5650, 'name': SearchDatastore_Task, 'duration_secs': 0.049261} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.534967] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a4bb264d-bb53-47e8-9428-53fd26e1e56e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.539968] env[63088]: DEBUG oslo_vmware.api [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Task: {'id': task-1284922, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.544184] env[63088]: DEBUG oslo_vmware.api [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Waiting for the task: (returnval){ [ 832.544184] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]5226653a-5298-1a7e-6d38-a81acc4b7691" [ 832.544184] env[63088]: _type = "Task" [ 832.544184] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.550437] env[63088]: DEBUG oslo_vmware.api [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]5226653a-5298-1a7e-6d38-a81acc4b7691, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.647144] env[63088]: DEBUG oslo_concurrency.lockutils [None req-da2e38c8-9951-4dcc-950e-b6d060a5ab7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.860s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 832.649521] env[63088]: DEBUG oslo_concurrency.lockutils [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.388s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 832.651497] env[63088]: INFO nova.compute.claims [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 00925f94-dbf9-453c-a124-d8434679aedf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 832.666918] env[63088]: INFO nova.scheduler.client.report [None req-da2e38c8-9951-4dcc-950e-b6d060a5ab7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Deleted allocations for instance 11ec987f-9908-4366-81ae-cbfe4de67b2d [ 832.742188] env[63088]: DEBUG oslo_concurrency.lockutils [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Acquiring lock "refresh_cache-1d421bf0-925d-408c-a612-b659da92799f" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 832.742188] env[63088]: DEBUG oslo_concurrency.lockutils [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Acquired lock "refresh_cache-1d421bf0-925d-408c-a612-b659da92799f" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 832.742188] env[63088]: DEBUG nova.network.neutron [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] [instance: 1d421bf0-925d-408c-a612-b659da92799f] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 832.743526] env[63088]: DEBUG nova.compute.manager [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: 58d561fe-26aa-4e94-8d55-cc70c361b479] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 832.770488] env[63088]: DEBUG nova.compute.manager [req-d6a45513-9aa3-4740-a1cb-76ec7148c55e req-4e4533f6-4f03-4477-8ac0-0b87f51cd0b9 service nova] [instance: 4f6e2d99-7300-42e6-8abf-362ddd4e8c6d] Received event network-vif-deleted-07c78e97-0b3c-45a7-859f-bbf74ce3c5f6 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 832.770752] env[63088]: DEBUG nova.compute.manager [req-d6a45513-9aa3-4740-a1cb-76ec7148c55e req-4e4533f6-4f03-4477-8ac0-0b87f51cd0b9 service nova] [instance: 1d421bf0-925d-408c-a612-b659da92799f] Received event network-vif-plugged-1c83da13-18b2-45c9-bc41-4d45e17606d5 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 832.770970] env[63088]: DEBUG oslo_concurrency.lockutils [req-d6a45513-9aa3-4740-a1cb-76ec7148c55e req-4e4533f6-4f03-4477-8ac0-0b87f51cd0b9 service nova] Acquiring lock "1d421bf0-925d-408c-a612-b659da92799f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 832.771190] env[63088]: DEBUG oslo_concurrency.lockutils [req-d6a45513-9aa3-4740-a1cb-76ec7148c55e req-4e4533f6-4f03-4477-8ac0-0b87f51cd0b9 service nova] Lock "1d421bf0-925d-408c-a612-b659da92799f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 832.771358] env[63088]: DEBUG oslo_concurrency.lockutils [req-d6a45513-9aa3-4740-a1cb-76ec7148c55e req-4e4533f6-4f03-4477-8ac0-0b87f51cd0b9 service nova] Lock "1d421bf0-925d-408c-a612-b659da92799f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 832.771520] env[63088]: DEBUG nova.compute.manager [req-d6a45513-9aa3-4740-a1cb-76ec7148c55e req-4e4533f6-4f03-4477-8ac0-0b87f51cd0b9 service nova] [instance: 1d421bf0-925d-408c-a612-b659da92799f] No waiting events found dispatching network-vif-plugged-1c83da13-18b2-45c9-bc41-4d45e17606d5 {{(pid=63088) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 832.771683] env[63088]: WARNING nova.compute.manager [req-d6a45513-9aa3-4740-a1cb-76ec7148c55e req-4e4533f6-4f03-4477-8ac0-0b87f51cd0b9 service nova] [instance: 1d421bf0-925d-408c-a612-b659da92799f] Received unexpected event network-vif-plugged-1c83da13-18b2-45c9-bc41-4d45e17606d5 for instance with vm_state building and task_state spawning. [ 832.771842] env[63088]: DEBUG nova.compute.manager [req-d6a45513-9aa3-4740-a1cb-76ec7148c55e req-4e4533f6-4f03-4477-8ac0-0b87f51cd0b9 service nova] [instance: 1d421bf0-925d-408c-a612-b659da92799f] Received event network-changed-1c83da13-18b2-45c9-bc41-4d45e17606d5 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 832.772035] env[63088]: DEBUG nova.compute.manager [req-d6a45513-9aa3-4740-a1cb-76ec7148c55e req-4e4533f6-4f03-4477-8ac0-0b87f51cd0b9 service nova] [instance: 1d421bf0-925d-408c-a612-b659da92799f] Refreshing instance network info cache due to event network-changed-1c83da13-18b2-45c9-bc41-4d45e17606d5. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 832.772167] env[63088]: DEBUG oslo_concurrency.lockutils [req-d6a45513-9aa3-4740-a1cb-76ec7148c55e req-4e4533f6-4f03-4477-8ac0-0b87f51cd0b9 service nova] Acquiring lock "refresh_cache-1d421bf0-925d-408c-a612-b659da92799f" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 832.807053] env[63088]: DEBUG nova.compute.manager [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] [instance: 2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 832.832269] env[63088]: DEBUG nova.virt.hardware [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 832.832567] env[63088]: DEBUG nova.virt.hardware [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 832.832725] env[63088]: DEBUG nova.virt.hardware [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 832.832950] env[63088]: DEBUG nova.virt.hardware [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 832.833127] env[63088]: DEBUG nova.virt.hardware [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 832.833280] env[63088]: DEBUG nova.virt.hardware [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 832.833492] env[63088]: DEBUG nova.virt.hardware [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 832.833650] env[63088]: DEBUG nova.virt.hardware [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 832.833819] env[63088]: DEBUG nova.virt.hardware [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 832.833981] env[63088]: DEBUG nova.virt.hardware [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 832.834225] env[63088]: DEBUG nova.virt.hardware [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 832.835144] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8430e4bc-1660-47a2-a204-3b2254c4e007 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.844328] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d630f75-2e25-4f72-a0e4-b838ab4b89d6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.890493] env[63088]: DEBUG oslo_vmware.api [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1284920, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07875} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.890925] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] Extended root virtual disk {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 832.891870] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58325457-f963-478d-b86b-829d7f96c351 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.917192] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] Reconfiguring VM instance instance-00000037 to attach disk [datastore1] 4847f15e-98d4-401c-91ff-067e84a85727/4847f15e-98d4-401c-91ff-067e84a85727.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 832.917890] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9cb1fb36-bc77-40a8-a906-7d3703617486 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.938658] env[63088]: DEBUG oslo_vmware.api [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Waiting for the task: (returnval){ [ 832.938658] env[63088]: value = "task-1284923" [ 832.938658] env[63088]: _type = "Task" [ 832.938658] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.947132] env[63088]: DEBUG oslo_vmware.api [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1284923, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.972946] env[63088]: INFO nova.compute.manager [None req-f844dd47-ebe5-483f-9e2f-773b32b15227 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Rescuing [ 832.973245] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f844dd47-ebe5-483f-9e2f-773b32b15227 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Acquiring lock "refresh_cache-07fd3e6f-e0af-467c-9039-238be5c58d25" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 832.973396] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f844dd47-ebe5-483f-9e2f-773b32b15227 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Acquired lock "refresh_cache-07fd3e6f-e0af-467c-9039-238be5c58d25" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 832.973562] env[63088]: DEBUG nova.network.neutron [None req-f844dd47-ebe5-483f-9e2f-773b32b15227 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 833.038409] env[63088]: DEBUG oslo_vmware.api [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Task: {'id': task-1284922, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.322951} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.038663] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 833.038846] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] [instance: b159649b-ebd5-4c7d-9074-dc0e6395a947] Deleted contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 833.039041] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] [instance: b159649b-ebd5-4c7d-9074-dc0e6395a947] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 833.052549] env[63088]: DEBUG oslo_vmware.api [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]5226653a-5298-1a7e-6d38-a81acc4b7691, 'name': SearchDatastore_Task, 'duration_secs': 0.013043} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.052796] env[63088]: DEBUG oslo_concurrency.lockutils [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 833.053055] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] 34ce411f-40c8-446e-b685-cd3ce07663dd/34ce411f-40c8-446e-b685-cd3ce07663dd.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 833.053300] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c849f64f-f422-4e5c-901a-48a90abc3a1d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.059856] env[63088]: DEBUG oslo_vmware.api [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Waiting for the task: (returnval){ [ 833.059856] env[63088]: value = "task-1284924" [ 833.059856] env[63088]: _type = "Task" [ 833.059856] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.067743] env[63088]: DEBUG oslo_vmware.api [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Task: {'id': task-1284924, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.174593] env[63088]: DEBUG oslo_concurrency.lockutils [None req-da2e38c8-9951-4dcc-950e-b6d060a5ab7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "11ec987f-9908-4366-81ae-cbfe4de67b2d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.045s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 833.266320] env[63088]: DEBUG oslo_concurrency.lockutils [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 833.295649] env[63088]: DEBUG nova.network.neutron [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] [instance: 1d421bf0-925d-408c-a612-b659da92799f] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 833.449585] env[63088]: DEBUG oslo_vmware.api [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1284923, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.532312] env[63088]: DEBUG nova.network.neutron [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] [instance: 1d421bf0-925d-408c-a612-b659da92799f] Updating instance_info_cache with network_info: [{"id": "1c83da13-18b2-45c9-bc41-4d45e17606d5", "address": "fa:16:3e:25:06:07", "network": {"id": "9e7064b1-65c0-4813-9423-e021c19e1041", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-2088266174-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "edf7d041b681480791590601dde8ba97", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3952eb02-1162-48ed-8227-9c138960d583", "external-id": "nsx-vlan-transportzone-250", "segmentation_id": 250, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1c83da13-18", "ovs_interfaceid": "1c83da13-18b2-45c9-bc41-4d45e17606d5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 833.578795] env[63088]: DEBUG oslo_vmware.api [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Task: {'id': task-1284924, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.615087] env[63088]: DEBUG nova.network.neutron [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] [instance: 2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68] Successfully updated port: 4c03d05a-bb4d-4d2b-bb2c-65f92a25278c {{(pid=63088) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 833.731363] env[63088]: DEBUG nova.network.neutron [None req-f844dd47-ebe5-483f-9e2f-773b32b15227 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Updating instance_info_cache with network_info: [{"id": "0c182796-7fb6-40ab-ba27-a43f0700dee1", "address": "fa:16:3e:63:25:2a", "network": {"id": "de116097-3737-4dcf-972c-88c49fa8995d", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-580707267-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "937eba1a40c84534b279d5798e234625", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3a80436-f7a9-431a-acec-aca3d76e3f9b", "external-id": "cl2-zone-339", "segmentation_id": 339, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c182796-7f", "ovs_interfaceid": "0c182796-7fb6-40ab-ba27-a43f0700dee1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 833.933785] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eaa41ae9-11e6-4923-a45a-decc1dd4eb93 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.944831] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-295d8dc0-4a88-4f31-ab7f-04c6e700a7aa {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.955302] env[63088]: DEBUG oslo_vmware.api [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1284923, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.980416] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c713332-ba16-4381-9d1c-dbda917cb813 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.989697] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc455c21-d520-4ed7-af61-882a2572d612 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.004082] env[63088]: DEBUG nova.compute.provider_tree [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 834.036238] env[63088]: DEBUG oslo_concurrency.lockutils [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Releasing lock "refresh_cache-1d421bf0-925d-408c-a612-b659da92799f" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 834.036575] env[63088]: DEBUG nova.compute.manager [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] [instance: 1d421bf0-925d-408c-a612-b659da92799f] Instance network_info: |[{"id": "1c83da13-18b2-45c9-bc41-4d45e17606d5", "address": "fa:16:3e:25:06:07", "network": {"id": "9e7064b1-65c0-4813-9423-e021c19e1041", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-2088266174-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "edf7d041b681480791590601dde8ba97", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3952eb02-1162-48ed-8227-9c138960d583", "external-id": "nsx-vlan-transportzone-250", "segmentation_id": 250, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1c83da13-18", "ovs_interfaceid": "1c83da13-18b2-45c9-bc41-4d45e17606d5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 834.036871] env[63088]: DEBUG oslo_concurrency.lockutils [req-d6a45513-9aa3-4740-a1cb-76ec7148c55e req-4e4533f6-4f03-4477-8ac0-0b87f51cd0b9 service nova] Acquired lock "refresh_cache-1d421bf0-925d-408c-a612-b659da92799f" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.037067] env[63088]: DEBUG nova.network.neutron [req-d6a45513-9aa3-4740-a1cb-76ec7148c55e req-4e4533f6-4f03-4477-8ac0-0b87f51cd0b9 service nova] [instance: 1d421bf0-925d-408c-a612-b659da92799f] Refreshing network info cache for port 1c83da13-18b2-45c9-bc41-4d45e17606d5 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 834.038533] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] [instance: 1d421bf0-925d-408c-a612-b659da92799f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:25:06:07', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3952eb02-1162-48ed-8227-9c138960d583', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1c83da13-18b2-45c9-bc41-4d45e17606d5', 'vif_model': 'vmxnet3'}] {{(pid=63088) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 834.048638] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Creating folder: Project (edf7d041b681480791590601dde8ba97). Parent ref: group-v275816. {{(pid=63088) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 834.053235] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5d7a4655-d64c-45b1-a67a-f9d2943f100f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.067634] env[63088]: INFO nova.virt.vmwareapi.vm_util [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Created folder: Project (edf7d041b681480791590601dde8ba97) in parent group-v275816. [ 834.067831] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Creating folder: Instances. Parent ref: group-v275861. {{(pid=63088) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 834.068470] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0fb51646-7f16-4d9c-ab3e-703b30f4f7c6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.075722] env[63088]: DEBUG oslo_vmware.api [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Task: {'id': task-1284924, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.013643} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.075975] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] 34ce411f-40c8-446e-b685-cd3ce07663dd/34ce411f-40c8-446e-b685-cd3ce07663dd.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 834.076203] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] [instance: 34ce411f-40c8-446e-b685-cd3ce07663dd] Extending root virtual disk to 1048576 {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 834.076459] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-30ecf467-96d3-450d-9aa0-7ff710e730f4 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.079777] env[63088]: INFO nova.virt.vmwareapi.vm_util [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Created folder: Instances in parent group-v275861. [ 834.080026] env[63088]: DEBUG oslo.service.loopingcall [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 834.080535] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1d421bf0-925d-408c-a612-b659da92799f] Creating VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 834.081247] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-95e20133-a0ca-4516-b57e-7d62314db3b3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.096939] env[63088]: DEBUG nova.virt.hardware [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 834.097198] env[63088]: DEBUG nova.virt.hardware [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 834.097356] env[63088]: DEBUG nova.virt.hardware [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 834.097541] env[63088]: DEBUG nova.virt.hardware [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 834.097682] env[63088]: DEBUG nova.virt.hardware [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 834.097827] env[63088]: DEBUG nova.virt.hardware [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 834.098043] env[63088]: DEBUG nova.virt.hardware [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 834.098207] env[63088]: DEBUG nova.virt.hardware [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 834.098372] env[63088]: DEBUG nova.virt.hardware [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 834.098532] env[63088]: DEBUG nova.virt.hardware [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 834.098707] env[63088]: DEBUG nova.virt.hardware [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 834.101051] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4af543a6-c6d9-4919-80f7-febe4ad2e8d7 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.103676] env[63088]: DEBUG oslo_vmware.api [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Waiting for the task: (returnval){ [ 834.103676] env[63088]: value = "task-1284927" [ 834.103676] env[63088]: _type = "Task" [ 834.103676] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.113246] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c1561c2-e863-49b6-a7c4-41de4c91189c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.116929] env[63088]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 834.116929] env[63088]: value = "task-1284928" [ 834.116929] env[63088]: _type = "Task" [ 834.116929] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.119065] env[63088]: DEBUG oslo_concurrency.lockutils [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Acquiring lock "refresh_cache-2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 834.119237] env[63088]: DEBUG oslo_concurrency.lockutils [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Acquired lock "refresh_cache-2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.119357] env[63088]: DEBUG nova.network.neutron [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] [instance: 2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 834.136150] env[63088]: DEBUG oslo_vmware.api [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Task: {'id': task-1284927, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.137371] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] [instance: b159649b-ebd5-4c7d-9074-dc0e6395a947] Instance VIF info [] {{(pid=63088) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 834.143091] env[63088]: DEBUG oslo.service.loopingcall [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 834.144152] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b159649b-ebd5-4c7d-9074-dc0e6395a947] Creating VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 834.144420] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-49cdb0fa-c057-4ddc-b003-bc4ff8bc560d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.161116] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1284928, 'name': CreateVM_Task} progress is 15%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.165975] env[63088]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 834.165975] env[63088]: value = "task-1284929" [ 834.165975] env[63088]: _type = "Task" [ 834.165975] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.173997] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1284929, 'name': CreateVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.234396] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f844dd47-ebe5-483f-9e2f-773b32b15227 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Releasing lock "refresh_cache-07fd3e6f-e0af-467c-9039-238be5c58d25" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 834.451074] env[63088]: DEBUG oslo_vmware.api [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1284923, 'name': ReconfigVM_Task, 'duration_secs': 1.186691} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.451074] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] Reconfigured VM instance instance-00000037 to attach disk [datastore1] 4847f15e-98d4-401c-91ff-067e84a85727/4847f15e-98d4-401c-91ff-067e84a85727.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 834.451584] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8ff12066-7f92-4c7b-b08b-c2c3278a3068 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.458177] env[63088]: DEBUG oslo_vmware.api [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Waiting for the task: (returnval){ [ 834.458177] env[63088]: value = "task-1284930" [ 834.458177] env[63088]: _type = "Task" [ 834.458177] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.466910] env[63088]: DEBUG oslo_vmware.api [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1284930, 'name': Rename_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.508095] env[63088]: DEBUG nova.scheduler.client.report [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 834.614547] env[63088]: DEBUG oslo_vmware.api [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Task: {'id': task-1284927, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07013} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.614842] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] [instance: 34ce411f-40c8-446e-b685-cd3ce07663dd] Extended root virtual disk {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 834.615695] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7e63f31-356e-4115-9453-68c207ae4cde {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.640453] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] [instance: 34ce411f-40c8-446e-b685-cd3ce07663dd] Reconfiguring VM instance instance-0000003c to attach disk [datastore1] 34ce411f-40c8-446e-b685-cd3ce07663dd/34ce411f-40c8-446e-b685-cd3ce07663dd.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 834.645016] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5879954f-894f-4b11-80a3-73d234fb4694 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.662649] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1284928, 'name': CreateVM_Task} progress is 99%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.670538] env[63088]: DEBUG oslo_vmware.api [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Waiting for the task: (returnval){ [ 834.670538] env[63088]: value = "task-1284931" [ 834.670538] env[63088]: _type = "Task" [ 834.670538] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.677661] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1284929, 'name': CreateVM_Task} progress is 25%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.684074] env[63088]: DEBUG oslo_vmware.api [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Task: {'id': task-1284931, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.689323] env[63088]: DEBUG nova.network.neutron [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] [instance: 2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 834.733199] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquiring lock "9d5f1cde-e787-4a2c-ac49-83ec135ff51c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 834.733444] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "9d5f1cde-e787-4a2c-ac49-83ec135ff51c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 834.766274] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-f844dd47-ebe5-483f-9e2f-773b32b15227 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 834.766531] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-27f43807-fd8f-4d00-b0b5-7d961448431d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.772986] env[63088]: DEBUG oslo_vmware.api [None req-f844dd47-ebe5-483f-9e2f-773b32b15227 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Waiting for the task: (returnval){ [ 834.772986] env[63088]: value = "task-1284932" [ 834.772986] env[63088]: _type = "Task" [ 834.772986] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.782651] env[63088]: DEBUG oslo_vmware.api [None req-f844dd47-ebe5-483f-9e2f-773b32b15227 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Task: {'id': task-1284932, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.859534] env[63088]: DEBUG nova.network.neutron [req-d6a45513-9aa3-4740-a1cb-76ec7148c55e req-4e4533f6-4f03-4477-8ac0-0b87f51cd0b9 service nova] [instance: 1d421bf0-925d-408c-a612-b659da92799f] Updated VIF entry in instance network info cache for port 1c83da13-18b2-45c9-bc41-4d45e17606d5. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 834.859983] env[63088]: DEBUG nova.network.neutron [req-d6a45513-9aa3-4740-a1cb-76ec7148c55e req-4e4533f6-4f03-4477-8ac0-0b87f51cd0b9 service nova] [instance: 1d421bf0-925d-408c-a612-b659da92799f] Updating instance_info_cache with network_info: [{"id": "1c83da13-18b2-45c9-bc41-4d45e17606d5", "address": "fa:16:3e:25:06:07", "network": {"id": "9e7064b1-65c0-4813-9423-e021c19e1041", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-2088266174-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "edf7d041b681480791590601dde8ba97", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3952eb02-1162-48ed-8227-9c138960d583", "external-id": "nsx-vlan-transportzone-250", "segmentation_id": 250, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1c83da13-18", "ovs_interfaceid": "1c83da13-18b2-45c9-bc41-4d45e17606d5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 834.888588] env[63088]: DEBUG nova.network.neutron [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] [instance: 2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68] Updating instance_info_cache with network_info: [{"id": "4c03d05a-bb4d-4d2b-bb2c-65f92a25278c", "address": "fa:16:3e:47:c7:b3", "network": {"id": "603ebdbc-21ee-430e-853e-4585b77f36c6", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-1660275285-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6bd87f64aa654aad8f632c4f22a2d266", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73eeba7c-29e1-4fdf-82b3-d62e63e86051", "external-id": "cl2-zone-659", "segmentation_id": 659, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c03d05a-bb", "ovs_interfaceid": "4c03d05a-bb4d-4d2b-bb2c-65f92a25278c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 834.912033] env[63088]: DEBUG nova.compute.manager [req-b20120e5-e36e-41a3-8d13-5e37194464bc req-3405b65e-2edf-47c7-acdc-e0a621ed7b61 service nova] [instance: 2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68] Received event network-vif-plugged-4c03d05a-bb4d-4d2b-bb2c-65f92a25278c {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 834.912033] env[63088]: DEBUG oslo_concurrency.lockutils [req-b20120e5-e36e-41a3-8d13-5e37194464bc req-3405b65e-2edf-47c7-acdc-e0a621ed7b61 service nova] Acquiring lock "2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 834.912033] env[63088]: DEBUG oslo_concurrency.lockutils [req-b20120e5-e36e-41a3-8d13-5e37194464bc req-3405b65e-2edf-47c7-acdc-e0a621ed7b61 service nova] Lock "2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 834.912270] env[63088]: DEBUG oslo_concurrency.lockutils [req-b20120e5-e36e-41a3-8d13-5e37194464bc req-3405b65e-2edf-47c7-acdc-e0a621ed7b61 service nova] Lock "2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 834.912470] env[63088]: DEBUG nova.compute.manager [req-b20120e5-e36e-41a3-8d13-5e37194464bc req-3405b65e-2edf-47c7-acdc-e0a621ed7b61 service nova] [instance: 2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68] No waiting events found dispatching network-vif-plugged-4c03d05a-bb4d-4d2b-bb2c-65f92a25278c {{(pid=63088) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 834.912806] env[63088]: WARNING nova.compute.manager [req-b20120e5-e36e-41a3-8d13-5e37194464bc req-3405b65e-2edf-47c7-acdc-e0a621ed7b61 service nova] [instance: 2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68] Received unexpected event network-vif-plugged-4c03d05a-bb4d-4d2b-bb2c-65f92a25278c for instance with vm_state building and task_state spawning. [ 834.912989] env[63088]: DEBUG nova.compute.manager [req-b20120e5-e36e-41a3-8d13-5e37194464bc req-3405b65e-2edf-47c7-acdc-e0a621ed7b61 service nova] [instance: 2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68] Received event network-changed-4c03d05a-bb4d-4d2b-bb2c-65f92a25278c {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 834.913213] env[63088]: DEBUG nova.compute.manager [req-b20120e5-e36e-41a3-8d13-5e37194464bc req-3405b65e-2edf-47c7-acdc-e0a621ed7b61 service nova] [instance: 2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68] Refreshing instance network info cache due to event network-changed-4c03d05a-bb4d-4d2b-bb2c-65f92a25278c. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 834.913417] env[63088]: DEBUG oslo_concurrency.lockutils [req-b20120e5-e36e-41a3-8d13-5e37194464bc req-3405b65e-2edf-47c7-acdc-e0a621ed7b61 service nova] Acquiring lock "refresh_cache-2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 834.967926] env[63088]: DEBUG oslo_vmware.api [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1284930, 'name': Rename_Task, 'duration_secs': 0.380311} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.968214] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 834.968456] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3bc7fe8a-32c5-49c8-8f3d-234c434963db {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.974639] env[63088]: DEBUG oslo_vmware.api [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Waiting for the task: (returnval){ [ 834.974639] env[63088]: value = "task-1284933" [ 834.974639] env[63088]: _type = "Task" [ 834.974639] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.982139] env[63088]: DEBUG oslo_vmware.api [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1284933, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.014320] env[63088]: DEBUG oslo_concurrency.lockutils [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.365s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 835.014856] env[63088]: DEBUG nova.compute.manager [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 00925f94-dbf9-453c-a124-d8434679aedf] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 835.017576] env[63088]: DEBUG oslo_concurrency.lockutils [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.779s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 835.018997] env[63088]: INFO nova.compute.claims [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 7f341d80-450a-47b8-b26d-15f2ce3e378c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 835.128400] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1284928, 'name': CreateVM_Task} progress is 99%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.178891] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1284929, 'name': CreateVM_Task, 'duration_secs': 0.706146} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.178891] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b159649b-ebd5-4c7d-9074-dc0e6395a947] Created VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 835.179143] env[63088]: DEBUG oslo_concurrency.lockutils [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 835.179313] env[63088]: DEBUG oslo_concurrency.lockutils [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 835.179611] env[63088]: DEBUG oslo_concurrency.lockutils [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 835.179876] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-760b1ec3-f662-42af-9091-e1e353d1cf7a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.185317] env[63088]: DEBUG oslo_vmware.api [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Task: {'id': task-1284931, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.188507] env[63088]: DEBUG oslo_vmware.api [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Waiting for the task: (returnval){ [ 835.188507] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52eafe31-4e69-495a-8d91-bb957cfaf70d" [ 835.188507] env[63088]: _type = "Task" [ 835.188507] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.195721] env[63088]: DEBUG oslo_vmware.api [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52eafe31-4e69-495a-8d91-bb957cfaf70d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.282230] env[63088]: DEBUG oslo_vmware.api [None req-f844dd47-ebe5-483f-9e2f-773b32b15227 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Task: {'id': task-1284932, 'name': PowerOffVM_Task, 'duration_secs': 0.196881} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.282526] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-f844dd47-ebe5-483f-9e2f-773b32b15227 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 835.283983] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52fba0de-cf63-49be-a31c-87ab3a9ecdde {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.303361] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-353ed3cc-a28c-4af8-a512-7e20f6f00737 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.341307] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-f844dd47-ebe5-483f-9e2f-773b32b15227 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 835.341744] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9e392a0c-d0f1-4fb9-85a3-1a1a3bd557d9 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.349524] env[63088]: DEBUG oslo_vmware.api [None req-f844dd47-ebe5-483f-9e2f-773b32b15227 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Waiting for the task: (returnval){ [ 835.349524] env[63088]: value = "task-1284934" [ 835.349524] env[63088]: _type = "Task" [ 835.349524] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.365045] env[63088]: DEBUG oslo_concurrency.lockutils [req-d6a45513-9aa3-4740-a1cb-76ec7148c55e req-4e4533f6-4f03-4477-8ac0-0b87f51cd0b9 service nova] Releasing lock "refresh_cache-1d421bf0-925d-408c-a612-b659da92799f" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 835.365936] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-f844dd47-ebe5-483f-9e2f-773b32b15227 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] VM already powered off {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 835.366076] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-f844dd47-ebe5-483f-9e2f-773b32b15227 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Processing image 1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 835.366448] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f844dd47-ebe5-483f-9e2f-773b32b15227 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 835.366716] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f844dd47-ebe5-483f-9e2f-773b32b15227 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 835.367067] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-f844dd47-ebe5-483f-9e2f-773b32b15227 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 835.367415] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4518cd25-165b-4a6f-8eea-9022bb092f53 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.386080] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-f844dd47-ebe5-483f-9e2f-773b32b15227 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 835.386080] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-f844dd47-ebe5-483f-9e2f-773b32b15227 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 835.386725] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-29256d25-cf49-4f68-9e9c-883d88fd2abe {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.392035] env[63088]: DEBUG oslo_concurrency.lockutils [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Releasing lock "refresh_cache-2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 835.392357] env[63088]: DEBUG nova.compute.manager [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] [instance: 2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68] Instance network_info: |[{"id": "4c03d05a-bb4d-4d2b-bb2c-65f92a25278c", "address": "fa:16:3e:47:c7:b3", "network": {"id": "603ebdbc-21ee-430e-853e-4585b77f36c6", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-1660275285-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6bd87f64aa654aad8f632c4f22a2d266", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73eeba7c-29e1-4fdf-82b3-d62e63e86051", "external-id": "cl2-zone-659", "segmentation_id": 659, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c03d05a-bb", "ovs_interfaceid": "4c03d05a-bb4d-4d2b-bb2c-65f92a25278c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 835.392701] env[63088]: DEBUG oslo_vmware.api [None req-f844dd47-ebe5-483f-9e2f-773b32b15227 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Waiting for the task: (returnval){ [ 835.392701] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52f3b4b1-f65a-0d12-fabe-d14a18a51826" [ 835.392701] env[63088]: _type = "Task" [ 835.392701] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.392945] env[63088]: DEBUG oslo_concurrency.lockutils [req-b20120e5-e36e-41a3-8d13-5e37194464bc req-3405b65e-2edf-47c7-acdc-e0a621ed7b61 service nova] Acquired lock "refresh_cache-2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 835.393169] env[63088]: DEBUG nova.network.neutron [req-b20120e5-e36e-41a3-8d13-5e37194464bc req-3405b65e-2edf-47c7-acdc-e0a621ed7b61 service nova] [instance: 2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68] Refreshing network info cache for port 4c03d05a-bb4d-4d2b-bb2c-65f92a25278c {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 835.395068] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] [instance: 2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:47:c7:b3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '73eeba7c-29e1-4fdf-82b3-d62e63e86051', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4c03d05a-bb4d-4d2b-bb2c-65f92a25278c', 'vif_model': 'vmxnet3'}] {{(pid=63088) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 835.404147] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Creating folder: Project (6bd87f64aa654aad8f632c4f22a2d266). Parent ref: group-v275816. {{(pid=63088) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 835.404147] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1cb6b28b-a008-4dfc-b254-15ff26433998 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.416472] env[63088]: DEBUG oslo_vmware.api [None req-f844dd47-ebe5-483f-9e2f-773b32b15227 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52f3b4b1-f65a-0d12-fabe-d14a18a51826, 'name': SearchDatastore_Task, 'duration_secs': 0.012271} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.418250] env[63088]: INFO nova.virt.vmwareapi.vm_util [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Created folder: Project (6bd87f64aa654aad8f632c4f22a2d266) in parent group-v275816. [ 835.418432] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Creating folder: Instances. Parent ref: group-v275865. {{(pid=63088) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 835.418660] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-69753221-02a6-4727-9c99-e3c2ddd23f39 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.421062] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-154c4f34-2e67-4b8f-8adb-fb838f8f5d5d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.425539] env[63088]: DEBUG oslo_vmware.api [None req-f844dd47-ebe5-483f-9e2f-773b32b15227 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Waiting for the task: (returnval){ [ 835.425539] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52e98714-ab19-ad85-3257-29785327cf44" [ 835.425539] env[63088]: _type = "Task" [ 835.425539] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.431744] env[63088]: INFO nova.virt.vmwareapi.vm_util [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Created folder: Instances in parent group-v275865. [ 835.432071] env[63088]: DEBUG oslo.service.loopingcall [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 835.434981] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68] Creating VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 835.435621] env[63088]: DEBUG oslo_vmware.api [None req-f844dd47-ebe5-483f-9e2f-773b32b15227 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52e98714-ab19-ad85-3257-29785327cf44, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.435836] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b8c5dce2-d2a4-46a2-9259-7129f50af19b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.455743] env[63088]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 835.455743] env[63088]: value = "task-1284937" [ 835.455743] env[63088]: _type = "Task" [ 835.455743] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.463660] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1284937, 'name': CreateVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.486443] env[63088]: DEBUG oslo_vmware.api [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1284933, 'name': PowerOnVM_Task, 'duration_secs': 0.433512} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.486854] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 835.487115] env[63088]: DEBUG nova.compute.manager [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 835.488262] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21c8ccea-d3d2-439d-8904-fb51f4b36ba6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.524371] env[63088]: DEBUG nova.compute.utils [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 835.529451] env[63088]: DEBUG nova.compute.manager [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 00925f94-dbf9-453c-a124-d8434679aedf] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 835.529752] env[63088]: DEBUG nova.network.neutron [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 00925f94-dbf9-453c-a124-d8434679aedf] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 835.577321] env[63088]: DEBUG nova.policy [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '37d241cdf58e4aa7b05f1f24a3c78296', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e27be174c0b3400193f9ae64a242da1a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 835.631647] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1284928, 'name': CreateVM_Task, 'duration_secs': 1.279048} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.631838] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1d421bf0-925d-408c-a612-b659da92799f] Created VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 835.632541] env[63088]: DEBUG oslo_concurrency.lockutils [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 835.632704] env[63088]: DEBUG oslo_concurrency.lockutils [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 835.633031] env[63088]: DEBUG oslo_concurrency.lockutils [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 835.633293] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-22ee65db-3bc7-4cc9-9759-cb96c22c4bd8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.637855] env[63088]: DEBUG oslo_vmware.api [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Waiting for the task: (returnval){ [ 835.637855] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52f86b88-41b6-a6d1-5cb2-24a3c4e62261" [ 835.637855] env[63088]: _type = "Task" [ 835.637855] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.645524] env[63088]: DEBUG oslo_vmware.api [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52f86b88-41b6-a6d1-5cb2-24a3c4e62261, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.680837] env[63088]: DEBUG oslo_vmware.api [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Task: {'id': task-1284931, 'name': ReconfigVM_Task, 'duration_secs': 0.934085} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.681190] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] [instance: 34ce411f-40c8-446e-b685-cd3ce07663dd] Reconfigured VM instance instance-0000003c to attach disk [datastore1] 34ce411f-40c8-446e-b685-cd3ce07663dd/34ce411f-40c8-446e-b685-cd3ce07663dd.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 835.681787] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3bfb32f9-9261-4edf-9ab4-c4f53decf755 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.687294] env[63088]: DEBUG oslo_vmware.api [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Waiting for the task: (returnval){ [ 835.687294] env[63088]: value = "task-1284938" [ 835.687294] env[63088]: _type = "Task" [ 835.687294] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.697931] env[63088]: DEBUG oslo_vmware.api [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Task: {'id': task-1284938, 'name': Rename_Task} progress is 6%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.701514] env[63088]: DEBUG oslo_vmware.api [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52eafe31-4e69-495a-8d91-bb957cfaf70d, 'name': SearchDatastore_Task, 'duration_secs': 0.014966} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.701813] env[63088]: DEBUG oslo_concurrency.lockutils [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 835.702092] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] [instance: b159649b-ebd5-4c7d-9074-dc0e6395a947] Processing image 1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 835.702318] env[63088]: DEBUG oslo_concurrency.lockutils [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 835.883714] env[63088]: DEBUG nova.network.neutron [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 00925f94-dbf9-453c-a124-d8434679aedf] Successfully created port: e2779ab6-aee9-4245-aab0-9d598a612844 {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 835.937997] env[63088]: DEBUG oslo_vmware.api [None req-f844dd47-ebe5-483f-9e2f-773b32b15227 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52e98714-ab19-ad85-3257-29785327cf44, 'name': SearchDatastore_Task, 'duration_secs': 0.015345} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.940544] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f844dd47-ebe5-483f-9e2f-773b32b15227 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 835.940847] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-f844dd47-ebe5-483f-9e2f-773b32b15227 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore2] 07fd3e6f-e0af-467c-9039-238be5c58d25/1e8c5d18-0a03-4e18-afe1-de5a6e255953-rescue.vmdk. {{(pid=63088) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 835.941181] env[63088]: DEBUG oslo_concurrency.lockutils [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 835.941390] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 835.941604] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-43dd7caf-dbd1-405a-8454-fb38f167ed2c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.944050] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-91b8cd8a-8578-4fc9-aacc-121571d3b20a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.950865] env[63088]: DEBUG oslo_vmware.api [None req-f844dd47-ebe5-483f-9e2f-773b32b15227 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Waiting for the task: (returnval){ [ 835.950865] env[63088]: value = "task-1284939" [ 835.950865] env[63088]: _type = "Task" [ 835.950865] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.954826] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 835.955011] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 835.957159] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-faa53707-5847-4f1d-a644-4f3c6ee788d3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.969273] env[63088]: DEBUG oslo_vmware.api [None req-f844dd47-ebe5-483f-9e2f-773b32b15227 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Task: {'id': task-1284939, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.973185] env[63088]: DEBUG oslo_vmware.api [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Waiting for the task: (returnval){ [ 835.973185] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]526d8185-933a-8336-1930-223d29648e53" [ 835.973185] env[63088]: _type = "Task" [ 835.973185] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.975931] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1284937, 'name': CreateVM_Task} progress is 99%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.984148] env[63088]: DEBUG oslo_vmware.api [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]526d8185-933a-8336-1930-223d29648e53, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.006182] env[63088]: DEBUG oslo_concurrency.lockutils [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 836.030164] env[63088]: DEBUG nova.compute.manager [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 00925f94-dbf9-453c-a124-d8434679aedf] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 836.151180] env[63088]: DEBUG oslo_vmware.api [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52f86b88-41b6-a6d1-5cb2-24a3c4e62261, 'name': SearchDatastore_Task, 'duration_secs': 0.043161} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.154190] env[63088]: DEBUG oslo_concurrency.lockutils [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 836.154702] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] [instance: 1d421bf0-925d-408c-a612-b659da92799f] Processing image 1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 836.154702] env[63088]: DEBUG oslo_concurrency.lockutils [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 836.154825] env[63088]: DEBUG oslo_concurrency.lockutils [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 836.155011] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 836.155654] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9f19f20f-6564-4455-8abb-c542964c3b2c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.166155] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 836.166155] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 836.166155] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8b390dd9-2081-408e-831c-09c6150993a5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.173966] env[63088]: DEBUG oslo_vmware.api [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Waiting for the task: (returnval){ [ 836.173966] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52352aa7-184a-4e4f-84cf-c3b40137074d" [ 836.173966] env[63088]: _type = "Task" [ 836.173966] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.181750] env[63088]: DEBUG oslo_vmware.api [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52352aa7-184a-4e4f-84cf-c3b40137074d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.198580] env[63088]: DEBUG oslo_vmware.api [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Task: {'id': task-1284938, 'name': Rename_Task, 'duration_secs': 0.135228} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.201296] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] [instance: 34ce411f-40c8-446e-b685-cd3ce07663dd] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 836.202641] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-577588bd-3bde-439b-8870-a19f7f572680 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.208263] env[63088]: DEBUG oslo_vmware.api [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Waiting for the task: (returnval){ [ 836.208263] env[63088]: value = "task-1284940" [ 836.208263] env[63088]: _type = "Task" [ 836.208263] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.216304] env[63088]: DEBUG oslo_vmware.api [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Task: {'id': task-1284940, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.401033] env[63088]: DEBUG nova.network.neutron [req-b20120e5-e36e-41a3-8d13-5e37194464bc req-3405b65e-2edf-47c7-acdc-e0a621ed7b61 service nova] [instance: 2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68] Updated VIF entry in instance network info cache for port 4c03d05a-bb4d-4d2b-bb2c-65f92a25278c. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 836.401311] env[63088]: DEBUG nova.network.neutron [req-b20120e5-e36e-41a3-8d13-5e37194464bc req-3405b65e-2edf-47c7-acdc-e0a621ed7b61 service nova] [instance: 2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68] Updating instance_info_cache with network_info: [{"id": "4c03d05a-bb4d-4d2b-bb2c-65f92a25278c", "address": "fa:16:3e:47:c7:b3", "network": {"id": "603ebdbc-21ee-430e-853e-4585b77f36c6", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-1660275285-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6bd87f64aa654aad8f632c4f22a2d266", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73eeba7c-29e1-4fdf-82b3-d62e63e86051", "external-id": "cl2-zone-659", "segmentation_id": 659, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c03d05a-bb", "ovs_interfaceid": "4c03d05a-bb4d-4d2b-bb2c-65f92a25278c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 836.424055] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0de97a8-dafc-48e4-a6ff-168e35af03ce {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.432520] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-034ba32d-956f-4d61-bdb1-1f5ca4a88a40 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.472302] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b8303d9-b432-4c9a-8ec7-f1eb2de26819 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.487874] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1284937, 'name': CreateVM_Task, 'duration_secs': 0.540979} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.488254] env[63088]: DEBUG oslo_vmware.api [None req-f844dd47-ebe5-483f-9e2f-773b32b15227 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Task: {'id': task-1284939, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.489517] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68] Created VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 836.490821] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62b70d07-ed61-4c5e-9670-9d3ce581b030 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.497265] env[63088]: DEBUG oslo_concurrency.lockutils [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 836.497513] env[63088]: DEBUG oslo_concurrency.lockutils [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 836.497841] env[63088]: DEBUG oslo_concurrency.lockutils [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 836.500016] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-582ce3ce-bf55-4adf-b678-176b88ea72af {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.508021] env[63088]: DEBUG oslo_vmware.api [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]526d8185-933a-8336-1930-223d29648e53, 'name': SearchDatastore_Task, 'duration_secs': 0.011664} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.514079] env[63088]: DEBUG nova.compute.provider_tree [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 836.516544] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3e36fbd0-73e5-44c7-8f60-729268c8b982 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.519048] env[63088]: DEBUG oslo_vmware.api [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Waiting for the task: (returnval){ [ 836.519048] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]528a705b-3c85-85c9-4d50-1c28a19ea9df" [ 836.519048] env[63088]: _type = "Task" [ 836.519048] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.523922] env[63088]: DEBUG oslo_vmware.api [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Waiting for the task: (returnval){ [ 836.523922] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]529f4327-4488-82ae-5a93-de8c153af20d" [ 836.523922] env[63088]: _type = "Task" [ 836.523922] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.531335] env[63088]: DEBUG oslo_vmware.api [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]528a705b-3c85-85c9-4d50-1c28a19ea9df, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.540621] env[63088]: DEBUG oslo_vmware.api [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]529f4327-4488-82ae-5a93-de8c153af20d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.685173] env[63088]: DEBUG oslo_vmware.api [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52352aa7-184a-4e4f-84cf-c3b40137074d, 'name': SearchDatastore_Task, 'duration_secs': 0.009804} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.686025] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3ee7d732-5443-4b1b-b20c-f56a300a95ee {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.692794] env[63088]: DEBUG oslo_vmware.api [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Waiting for the task: (returnval){ [ 836.692794] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52048e4d-eaec-8c56-6eb6-7990157daccd" [ 836.692794] env[63088]: _type = "Task" [ 836.692794] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.701461] env[63088]: DEBUG oslo_vmware.api [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52048e4d-eaec-8c56-6eb6-7990157daccd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.719082] env[63088]: DEBUG oslo_vmware.api [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Task: {'id': task-1284940, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.847956] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0d443006-c001-488d-9c7f-f5cbe25ea71a tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquiring lock "4847f15e-98d4-401c-91ff-067e84a85727" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 836.847956] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0d443006-c001-488d-9c7f-f5cbe25ea71a tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Lock "4847f15e-98d4-401c-91ff-067e84a85727" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 836.847956] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0d443006-c001-488d-9c7f-f5cbe25ea71a tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquiring lock "4847f15e-98d4-401c-91ff-067e84a85727-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 836.848176] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0d443006-c001-488d-9c7f-f5cbe25ea71a tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Lock "4847f15e-98d4-401c-91ff-067e84a85727-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 836.849098] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0d443006-c001-488d-9c7f-f5cbe25ea71a tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Lock "4847f15e-98d4-401c-91ff-067e84a85727-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 836.851775] env[63088]: INFO nova.compute.manager [None req-0d443006-c001-488d-9c7f-f5cbe25ea71a tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] Terminating instance [ 836.854044] env[63088]: DEBUG nova.compute.manager [None req-0d443006-c001-488d-9c7f-f5cbe25ea71a tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 836.854249] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-0d443006-c001-488d-9c7f-f5cbe25ea71a tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 836.855079] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b46f356c-e198-475a-bbb6-29609bb29211 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.863044] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d443006-c001-488d-9c7f-f5cbe25ea71a tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 836.863351] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c5a566ba-e988-4fce-be94-ebe4840c9664 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.870038] env[63088]: DEBUG oslo_vmware.api [None req-0d443006-c001-488d-9c7f-f5cbe25ea71a tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Waiting for the task: (returnval){ [ 836.870038] env[63088]: value = "task-1284941" [ 836.870038] env[63088]: _type = "Task" [ 836.870038] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.879571] env[63088]: DEBUG oslo_vmware.api [None req-0d443006-c001-488d-9c7f-f5cbe25ea71a tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1284941, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.911120] env[63088]: DEBUG oslo_concurrency.lockutils [req-b20120e5-e36e-41a3-8d13-5e37194464bc req-3405b65e-2edf-47c7-acdc-e0a621ed7b61 service nova] Releasing lock "refresh_cache-2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 836.975360] env[63088]: DEBUG oslo_vmware.api [None req-f844dd47-ebe5-483f-9e2f-773b32b15227 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Task: {'id': task-1284939, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.817036} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.975653] env[63088]: INFO nova.virt.vmwareapi.ds_util [None req-f844dd47-ebe5-483f-9e2f-773b32b15227 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore2] 07fd3e6f-e0af-467c-9039-238be5c58d25/1e8c5d18-0a03-4e18-afe1-de5a6e255953-rescue.vmdk. [ 836.976513] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8260a9ea-b504-4556-8558-9c7bdf6ddbd5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.002909] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-f844dd47-ebe5-483f-9e2f-773b32b15227 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Reconfiguring VM instance instance-0000003a to attach disk [datastore2] 07fd3e6f-e0af-467c-9039-238be5c58d25/1e8c5d18-0a03-4e18-afe1-de5a6e255953-rescue.vmdk or device None with type thin {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 837.003190] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1f657e9f-669f-4269-aea1-d64765a06d57 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.020445] env[63088]: DEBUG nova.scheduler.client.report [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 837.027713] env[63088]: DEBUG oslo_vmware.api [None req-f844dd47-ebe5-483f-9e2f-773b32b15227 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Waiting for the task: (returnval){ [ 837.027713] env[63088]: value = "task-1284942" [ 837.027713] env[63088]: _type = "Task" [ 837.027713] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.038173] env[63088]: DEBUG oslo_vmware.api [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]528a705b-3c85-85c9-4d50-1c28a19ea9df, 'name': SearchDatastore_Task, 'duration_secs': 0.061902} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.039081] env[63088]: DEBUG oslo_concurrency.lockutils [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 837.039322] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] [instance: 2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68] Processing image 1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 837.039528] env[63088]: DEBUG oslo_concurrency.lockutils [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 837.045622] env[63088]: DEBUG nova.compute.manager [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 00925f94-dbf9-453c-a124-d8434679aedf] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 837.047701] env[63088]: DEBUG oslo_vmware.api [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]529f4327-4488-82ae-5a93-de8c153af20d, 'name': SearchDatastore_Task, 'duration_secs': 0.065856} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.051174] env[63088]: DEBUG oslo_concurrency.lockutils [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 837.051428] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore2] b159649b-ebd5-4c7d-9074-dc0e6395a947/b159649b-ebd5-4c7d-9074-dc0e6395a947.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 837.052014] env[63088]: DEBUG oslo_vmware.api [None req-f844dd47-ebe5-483f-9e2f-773b32b15227 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Task: {'id': task-1284942, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.052304] env[63088]: DEBUG oslo_concurrency.lockutils [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.052513] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 837.052867] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8c0f7954-e7c2-45ec-8c17-f44e5b4680bc {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.055649] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2ca34335-f10c-4837-a0c8-262d942356ab {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.063134] env[63088]: DEBUG oslo_vmware.api [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Waiting for the task: (returnval){ [ 837.063134] env[63088]: value = "task-1284943" [ 837.063134] env[63088]: _type = "Task" [ 837.063134] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.068985] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 837.069189] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 837.071804] env[63088]: DEBUG nova.virt.hardware [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 837.072038] env[63088]: DEBUG nova.virt.hardware [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 837.072202] env[63088]: DEBUG nova.virt.hardware [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 837.072384] env[63088]: DEBUG nova.virt.hardware [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 837.072530] env[63088]: DEBUG nova.virt.hardware [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 837.072675] env[63088]: DEBUG nova.virt.hardware [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 837.072891] env[63088]: DEBUG nova.virt.hardware [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 837.073063] env[63088]: DEBUG nova.virt.hardware [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 837.073233] env[63088]: DEBUG nova.virt.hardware [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 837.073396] env[63088]: DEBUG nova.virt.hardware [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 837.073634] env[63088]: DEBUG nova.virt.hardware [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 837.074210] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0df74062-212e-4d23-9920-35e8822e0d97 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.076765] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-400ec418-a783-497f-9d2a-1ec147f85e40 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.082158] env[63088]: DEBUG oslo_vmware.api [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Task: {'id': task-1284943, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.085144] env[63088]: DEBUG oslo_vmware.api [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Waiting for the task: (returnval){ [ 837.085144] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52064c67-e117-401b-3b17-b64e2ea935bd" [ 837.085144] env[63088]: _type = "Task" [ 837.085144] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.090525] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbc7c8a3-120c-496c-be07-afdb6416fac7 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.099969] env[63088]: DEBUG oslo_vmware.api [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52064c67-e117-401b-3b17-b64e2ea935bd, 'name': SearchDatastore_Task, 'duration_secs': 0.009861} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.108591] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-087a461d-1cbe-498c-b4b4-61952657dd50 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.113940] env[63088]: DEBUG oslo_vmware.api [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Waiting for the task: (returnval){ [ 837.113940] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]520b3036-0651-6bf9-59ed-e8d08b419d3d" [ 837.113940] env[63088]: _type = "Task" [ 837.113940] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.121852] env[63088]: DEBUG oslo_vmware.api [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]520b3036-0651-6bf9-59ed-e8d08b419d3d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.203054] env[63088]: DEBUG oslo_vmware.api [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52048e4d-eaec-8c56-6eb6-7990157daccd, 'name': SearchDatastore_Task, 'duration_secs': 0.022021} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.203337] env[63088]: DEBUG oslo_concurrency.lockutils [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 837.203642] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] 1d421bf0-925d-408c-a612-b659da92799f/1d421bf0-925d-408c-a612-b659da92799f.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 837.203934] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2c2bed02-de1f-4cc8-8bcd-bf21673b02f8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.214043] env[63088]: DEBUG oslo_vmware.api [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Waiting for the task: (returnval){ [ 837.214043] env[63088]: value = "task-1284944" [ 837.214043] env[63088]: _type = "Task" [ 837.214043] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.219794] env[63088]: DEBUG oslo_vmware.api [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Task: {'id': task-1284940, 'name': PowerOnVM_Task, 'duration_secs': 0.543747} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.220361] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] [instance: 34ce411f-40c8-446e-b685-cd3ce07663dd] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 837.220566] env[63088]: INFO nova.compute.manager [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] [instance: 34ce411f-40c8-446e-b685-cd3ce07663dd] Took 9.16 seconds to spawn the instance on the hypervisor. [ 837.220789] env[63088]: DEBUG nova.compute.manager [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] [instance: 34ce411f-40c8-446e-b685-cd3ce07663dd] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 837.221526] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2389479e-1e43-4bab-a586-0cbc913ea6c7 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.226814] env[63088]: DEBUG oslo_vmware.api [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Task: {'id': task-1284944, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.381102] env[63088]: DEBUG oslo_vmware.api [None req-0d443006-c001-488d-9c7f-f5cbe25ea71a tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1284941, 'name': PowerOffVM_Task, 'duration_secs': 0.194102} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.381797] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d443006-c001-488d-9c7f-f5cbe25ea71a tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 837.382182] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-0d443006-c001-488d-9c7f-f5cbe25ea71a tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] Unregistering the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 837.382488] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f2996977-40e7-49fd-a7a1-caeef583565c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.448989] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-0d443006-c001-488d-9c7f-f5cbe25ea71a tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] Unregistered the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 837.449342] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-0d443006-c001-488d-9c7f-f5cbe25ea71a tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] Deleting contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 837.449457] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-0d443006-c001-488d-9c7f-f5cbe25ea71a tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Deleting the datastore file [datastore1] 4847f15e-98d4-401c-91ff-067e84a85727 {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 837.449821] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e54b58b4-0166-46d8-83e9-9bbf8ba080cd {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.460625] env[63088]: DEBUG oslo_vmware.api [None req-0d443006-c001-488d-9c7f-f5cbe25ea71a tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Waiting for the task: (returnval){ [ 837.460625] env[63088]: value = "task-1284946" [ 837.460625] env[63088]: _type = "Task" [ 837.460625] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.469046] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquiring lock "2269e520-08b6-433f-8fe9-7b84d91e02d7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 837.469238] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Lock "2269e520-08b6-433f-8fe9-7b84d91e02d7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 837.475701] env[63088]: DEBUG oslo_vmware.api [None req-0d443006-c001-488d-9c7f-f5cbe25ea71a tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1284946, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.491308] env[63088]: DEBUG nova.compute.manager [req-b649b624-a1e7-4870-8628-9821ff52f0ca req-2d262c7d-5723-4dcb-9df7-d4fbbd98ce9b service nova] [instance: 00925f94-dbf9-453c-a124-d8434679aedf] Received event network-vif-plugged-e2779ab6-aee9-4245-aab0-9d598a612844 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 837.491564] env[63088]: DEBUG oslo_concurrency.lockutils [req-b649b624-a1e7-4870-8628-9821ff52f0ca req-2d262c7d-5723-4dcb-9df7-d4fbbd98ce9b service nova] Acquiring lock "00925f94-dbf9-453c-a124-d8434679aedf-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 837.491917] env[63088]: DEBUG oslo_concurrency.lockutils [req-b649b624-a1e7-4870-8628-9821ff52f0ca req-2d262c7d-5723-4dcb-9df7-d4fbbd98ce9b service nova] Lock "00925f94-dbf9-453c-a124-d8434679aedf-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 837.491917] env[63088]: DEBUG oslo_concurrency.lockutils [req-b649b624-a1e7-4870-8628-9821ff52f0ca req-2d262c7d-5723-4dcb-9df7-d4fbbd98ce9b service nova] Lock "00925f94-dbf9-453c-a124-d8434679aedf-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 837.492128] env[63088]: DEBUG nova.compute.manager [req-b649b624-a1e7-4870-8628-9821ff52f0ca req-2d262c7d-5723-4dcb-9df7-d4fbbd98ce9b service nova] [instance: 00925f94-dbf9-453c-a124-d8434679aedf] No waiting events found dispatching network-vif-plugged-e2779ab6-aee9-4245-aab0-9d598a612844 {{(pid=63088) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 837.492375] env[63088]: WARNING nova.compute.manager [req-b649b624-a1e7-4870-8628-9821ff52f0ca req-2d262c7d-5723-4dcb-9df7-d4fbbd98ce9b service nova] [instance: 00925f94-dbf9-453c-a124-d8434679aedf] Received unexpected event network-vif-plugged-e2779ab6-aee9-4245-aab0-9d598a612844 for instance with vm_state building and task_state spawning. [ 837.529863] env[63088]: DEBUG oslo_concurrency.lockutils [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.512s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 837.531120] env[63088]: DEBUG nova.compute.manager [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 7f341d80-450a-47b8-b26d-15f2ce3e378c] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 837.534908] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.501s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 837.536401] env[63088]: INFO nova.compute.claims [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 8b564e29-96cf-4abf-963d-142b413fb464] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 837.548889] env[63088]: DEBUG oslo_vmware.api [None req-f844dd47-ebe5-483f-9e2f-773b32b15227 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Task: {'id': task-1284942, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.576231] env[63088]: DEBUG oslo_vmware.api [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Task: {'id': task-1284943, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.605619] env[63088]: DEBUG nova.network.neutron [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 00925f94-dbf9-453c-a124-d8434679aedf] Successfully updated port: e2779ab6-aee9-4245-aab0-9d598a612844 {{(pid=63088) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 837.627694] env[63088]: DEBUG oslo_vmware.api [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]520b3036-0651-6bf9-59ed-e8d08b419d3d, 'name': SearchDatastore_Task, 'duration_secs': 0.030161} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.627982] env[63088]: DEBUG oslo_concurrency.lockutils [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 837.628276] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore2] 2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68/2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 837.628661] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9c5db901-337b-45e4-880d-b2c1d975b3a8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.636932] env[63088]: DEBUG oslo_vmware.api [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Waiting for the task: (returnval){ [ 837.636932] env[63088]: value = "task-1284947" [ 837.636932] env[63088]: _type = "Task" [ 837.636932] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.646192] env[63088]: DEBUG oslo_vmware.api [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Task: {'id': task-1284947, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.724996] env[63088]: DEBUG oslo_vmware.api [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Task: {'id': task-1284944, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.743553] env[63088]: INFO nova.compute.manager [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] [instance: 34ce411f-40c8-446e-b685-cd3ce07663dd] Took 30.47 seconds to build instance. [ 837.973397] env[63088]: DEBUG oslo_vmware.api [None req-0d443006-c001-488d-9c7f-f5cbe25ea71a tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1284946, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.044451] env[63088]: DEBUG nova.compute.utils [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 838.055658] env[63088]: DEBUG nova.compute.manager [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 7f341d80-450a-47b8-b26d-15f2ce3e378c] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 838.055996] env[63088]: DEBUG nova.network.neutron [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 7f341d80-450a-47b8-b26d-15f2ce3e378c] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 838.058954] env[63088]: DEBUG oslo_vmware.api [None req-f844dd47-ebe5-483f-9e2f-773b32b15227 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Task: {'id': task-1284942, 'name': ReconfigVM_Task, 'duration_secs': 0.650821} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.061176] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-f844dd47-ebe5-483f-9e2f-773b32b15227 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Reconfigured VM instance instance-0000003a to attach disk [datastore2] 07fd3e6f-e0af-467c-9039-238be5c58d25/1e8c5d18-0a03-4e18-afe1-de5a6e255953-rescue.vmdk or device None with type thin {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 838.062663] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f78c7d35-8a5a-4a2b-9b0c-353e105230fe {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.095623] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1874dc6b-7620-480b-9300-d2d09a73abd8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.116951] env[63088]: DEBUG oslo_concurrency.lockutils [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Acquiring lock "refresh_cache-00925f94-dbf9-453c-a124-d8434679aedf" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 838.117351] env[63088]: DEBUG oslo_concurrency.lockutils [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Acquired lock "refresh_cache-00925f94-dbf9-453c-a124-d8434679aedf" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.117536] env[63088]: DEBUG nova.network.neutron [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 00925f94-dbf9-453c-a124-d8434679aedf] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 838.119598] env[63088]: DEBUG oslo_vmware.api [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Task: {'id': task-1284943, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.536618} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.120595] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore2] b159649b-ebd5-4c7d-9074-dc0e6395a947/b159649b-ebd5-4c7d-9074-dc0e6395a947.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 838.120886] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] [instance: b159649b-ebd5-4c7d-9074-dc0e6395a947] Extending root virtual disk to 1048576 {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 838.121222] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-158947c2-93c6-4f5e-bb25-f712c76613c0 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.127284] env[63088]: DEBUG oslo_vmware.api [None req-f844dd47-ebe5-483f-9e2f-773b32b15227 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Waiting for the task: (returnval){ [ 838.127284] env[63088]: value = "task-1284948" [ 838.127284] env[63088]: _type = "Task" [ 838.127284] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.136092] env[63088]: DEBUG oslo_vmware.api [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Waiting for the task: (returnval){ [ 838.136092] env[63088]: value = "task-1284949" [ 838.136092] env[63088]: _type = "Task" [ 838.136092] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.143997] env[63088]: DEBUG oslo_vmware.api [None req-f844dd47-ebe5-483f-9e2f-773b32b15227 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Task: {'id': task-1284948, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.153278] env[63088]: DEBUG oslo_vmware.api [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Task: {'id': task-1284949, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.157675] env[63088]: DEBUG oslo_vmware.api [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Task: {'id': task-1284947, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.170025] env[63088]: DEBUG nova.policy [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '37d241cdf58e4aa7b05f1f24a3c78296', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e27be174c0b3400193f9ae64a242da1a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 838.225782] env[63088]: DEBUG oslo_vmware.api [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Task: {'id': task-1284944, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.669962} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.226054] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] 1d421bf0-925d-408c-a612-b659da92799f/1d421bf0-925d-408c-a612-b659da92799f.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 838.226263] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] [instance: 1d421bf0-925d-408c-a612-b659da92799f] Extending root virtual disk to 1048576 {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 838.226507] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f9db0f0e-c008-48db-bab1-ee1d76bad26b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.234214] env[63088]: DEBUG oslo_vmware.api [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Waiting for the task: (returnval){ [ 838.234214] env[63088]: value = "task-1284950" [ 838.234214] env[63088]: _type = "Task" [ 838.234214] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.241146] env[63088]: DEBUG oslo_vmware.api [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Task: {'id': task-1284950, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.245740] env[63088]: DEBUG oslo_concurrency.lockutils [None req-eb658cfc-a02d-4663-a967-c9d9bd3e2ace tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Lock "34ce411f-40c8-446e-b685-cd3ce07663dd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 66.227s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 838.471212] env[63088]: DEBUG oslo_vmware.api [None req-0d443006-c001-488d-9c7f-f5cbe25ea71a tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1284946, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.549597} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.471526] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-0d443006-c001-488d-9c7f-f5cbe25ea71a tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 838.471664] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-0d443006-c001-488d-9c7f-f5cbe25ea71a tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] Deleted contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 838.471807] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-0d443006-c001-488d-9c7f-f5cbe25ea71a tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 838.471978] env[63088]: INFO nova.compute.manager [None req-0d443006-c001-488d-9c7f-f5cbe25ea71a tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] Took 1.62 seconds to destroy the instance on the hypervisor. [ 838.472236] env[63088]: DEBUG oslo.service.loopingcall [None req-0d443006-c001-488d-9c7f-f5cbe25ea71a tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 838.472424] env[63088]: DEBUG nova.compute.manager [-] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 838.472518] env[63088]: DEBUG nova.network.neutron [-] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 838.549754] env[63088]: DEBUG nova.compute.manager [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 7f341d80-450a-47b8-b26d-15f2ce3e378c] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 838.644277] env[63088]: DEBUG oslo_vmware.api [None req-f844dd47-ebe5-483f-9e2f-773b32b15227 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Task: {'id': task-1284948, 'name': ReconfigVM_Task, 'duration_secs': 0.198406} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.644673] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-f844dd47-ebe5-483f-9e2f-773b32b15227 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 838.649693] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-52a041e0-5686-4836-9e4d-2c6b979b2cf2 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.657509] env[63088]: DEBUG oslo_vmware.api [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Task: {'id': task-1284949, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.081438} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.658993] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] [instance: b159649b-ebd5-4c7d-9074-dc0e6395a947] Extended root virtual disk {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 838.659791] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ec728c9-005f-46e6-967d-154a466482cb {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.666365] env[63088]: DEBUG oslo_vmware.api [None req-f844dd47-ebe5-483f-9e2f-773b32b15227 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Waiting for the task: (returnval){ [ 838.666365] env[63088]: value = "task-1284951" [ 838.666365] env[63088]: _type = "Task" [ 838.666365] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.666590] env[63088]: DEBUG oslo_vmware.api [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Task: {'id': task-1284947, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.587976} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.667208] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore2] 2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68/2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 838.667423] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] [instance: 2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68] Extending root virtual disk to 1048576 {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 838.673566] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e36488c7-4c3a-409c-8642-ec5a19b76a5e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.696341] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] [instance: b159649b-ebd5-4c7d-9074-dc0e6395a947] Reconfiguring VM instance instance-0000003b to attach disk [datastore2] b159649b-ebd5-4c7d-9074-dc0e6395a947/b159649b-ebd5-4c7d-9074-dc0e6395a947.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 838.697703] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-874d55e8-78f0-46c1-b836-462ced69874d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.724023] env[63088]: DEBUG oslo_vmware.api [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Waiting for the task: (returnval){ [ 838.724023] env[63088]: value = "task-1284952" [ 838.724023] env[63088]: _type = "Task" [ 838.724023] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.724023] env[63088]: DEBUG oslo_vmware.api [None req-f844dd47-ebe5-483f-9e2f-773b32b15227 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Task: {'id': task-1284951, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.730319] env[63088]: DEBUG oslo_vmware.api [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Waiting for the task: (returnval){ [ 838.730319] env[63088]: value = "task-1284953" [ 838.730319] env[63088]: _type = "Task" [ 838.730319] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.736533] env[63088]: DEBUG oslo_vmware.api [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Task: {'id': task-1284952, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.742801] env[63088]: DEBUG oslo_vmware.api [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Task: {'id': task-1284953, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.748442] env[63088]: DEBUG nova.compute.manager [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 9d5f1cde-e787-4a2c-ac49-83ec135ff51c] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 838.751042] env[63088]: DEBUG oslo_vmware.api [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Task: {'id': task-1284950, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064368} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.751504] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] [instance: 1d421bf0-925d-408c-a612-b659da92799f] Extended root virtual disk {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 838.752306] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31fe63ba-7659-4179-9d2b-d99d026f5bf4 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.777783] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] [instance: 1d421bf0-925d-408c-a612-b659da92799f] Reconfiguring VM instance instance-0000003d to attach disk [datastore1] 1d421bf0-925d-408c-a612-b659da92799f/1d421bf0-925d-408c-a612-b659da92799f.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 838.778121] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d2637f34-c791-4588-b4bf-18b8b2b97261 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.799988] env[63088]: DEBUG oslo_vmware.api [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Waiting for the task: (returnval){ [ 838.799988] env[63088]: value = "task-1284954" [ 838.799988] env[63088]: _type = "Task" [ 838.799988] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.810049] env[63088]: DEBUG oslo_vmware.api [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Task: {'id': task-1284954, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.830349] env[63088]: DEBUG nova.network.neutron [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 00925f94-dbf9-453c-a124-d8434679aedf] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 838.837878] env[63088]: DEBUG nova.compute.manager [req-3040807d-2c27-4e5d-8d6e-1957a4609e10 req-ce695102-2fec-4b8f-a3c6-4ef287592d49 service nova] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] Received event network-vif-deleted-53b4bd8d-df8b-4069-a75a-af46b605dddb {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 838.838173] env[63088]: INFO nova.compute.manager [req-3040807d-2c27-4e5d-8d6e-1957a4609e10 req-ce695102-2fec-4b8f-a3c6-4ef287592d49 service nova] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] Neutron deleted interface 53b4bd8d-df8b-4069-a75a-af46b605dddb; detaching it from the instance and deleting it from the info cache [ 838.838314] env[63088]: DEBUG nova.network.neutron [req-3040807d-2c27-4e5d-8d6e-1957a4609e10 req-ce695102-2fec-4b8f-a3c6-4ef287592d49 service nova] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 838.929307] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a464335-5076-46ae-80c7-6db18ae60ee2 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.939918] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b6e873b-17d8-4006-8e6a-a087238ead16 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.980823] env[63088]: DEBUG nova.network.neutron [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 7f341d80-450a-47b8-b26d-15f2ce3e378c] Successfully created port: 711f2731-680e-4328-97c8-0a2d1bbf79eb {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 838.986408] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e458db9-6f2c-48dd-9f0a-4f4b245e3cae {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.995434] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e9a6278-c57f-4739-a198-1507169aa1b1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.010888] env[63088]: DEBUG nova.compute.provider_tree [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 839.180854] env[63088]: DEBUG oslo_vmware.api [None req-f844dd47-ebe5-483f-9e2f-773b32b15227 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Task: {'id': task-1284951, 'name': PowerOnVM_Task, 'duration_secs': 0.414321} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.182095] env[63088]: DEBUG nova.network.neutron [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 00925f94-dbf9-453c-a124-d8434679aedf] Updating instance_info_cache with network_info: [{"id": "e2779ab6-aee9-4245-aab0-9d598a612844", "address": "fa:16:3e:da:19:08", "network": {"id": "79155637-c844-4012-a77a-d04f9ec9ab18", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-2144942540-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e27be174c0b3400193f9ae64a242da1a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape2779ab6-ae", "ovs_interfaceid": "e2779ab6-aee9-4245-aab0-9d598a612844", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 839.183736] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-f844dd47-ebe5-483f-9e2f-773b32b15227 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 839.190021] env[63088]: DEBUG nova.compute.manager [None req-f844dd47-ebe5-483f-9e2f-773b32b15227 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 839.190021] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e227b0a2-33b9-470f-a7a6-6de5646eb44e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.223022] env[63088]: DEBUG nova.network.neutron [-] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 839.235173] env[63088]: DEBUG oslo_vmware.api [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Task: {'id': task-1284952, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069172} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.237520] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] [instance: 2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68] Extended root virtual disk {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 839.237520] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5a7ad73-81b0-4004-99a2-b2cb08f28983 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.242425] env[63088]: DEBUG oslo_vmware.api [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Task: {'id': task-1284953, 'name': ReconfigVM_Task, 'duration_secs': 0.460653} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.243088] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] [instance: b159649b-ebd5-4c7d-9074-dc0e6395a947] Reconfigured VM instance instance-0000003b to attach disk [datastore2] b159649b-ebd5-4c7d-9074-dc0e6395a947/b159649b-ebd5-4c7d-9074-dc0e6395a947.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 839.243792] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-51848757-e4f1-4f7c-b511-54860cf6a9f7 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.264916] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] [instance: 2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68] Reconfiguring VM instance instance-0000003e to attach disk [datastore2] 2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68/2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 839.267878] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-816ba374-43f5-43b5-9a77-159cf1fffc39 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.284087] env[63088]: DEBUG oslo_vmware.api [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Waiting for the task: (returnval){ [ 839.284087] env[63088]: value = "task-1284955" [ 839.284087] env[63088]: _type = "Task" [ 839.284087] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.291077] env[63088]: DEBUG oslo_vmware.api [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Waiting for the task: (returnval){ [ 839.291077] env[63088]: value = "task-1284956" [ 839.291077] env[63088]: _type = "Task" [ 839.291077] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.297995] env[63088]: DEBUG oslo_vmware.api [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Task: {'id': task-1284955, 'name': Rename_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.299067] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 839.308315] env[63088]: DEBUG oslo_vmware.api [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Task: {'id': task-1284956, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.314683] env[63088]: DEBUG oslo_vmware.api [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Task: {'id': task-1284954, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.343254] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fcfa4ccb-fd07-4cff-82a5-2b1c3571e99f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.354059] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bc99d1e-e9d4-4dc5-9201-af4e15ef6c23 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.381262] env[63088]: DEBUG nova.compute.manager [req-3040807d-2c27-4e5d-8d6e-1957a4609e10 req-ce695102-2fec-4b8f-a3c6-4ef287592d49 service nova] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] Detach interface failed, port_id=53b4bd8d-df8b-4069-a75a-af46b605dddb, reason: Instance 4847f15e-98d4-401c-91ff-067e84a85727 could not be found. {{(pid=63088) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 839.514474] env[63088]: DEBUG nova.scheduler.client.report [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 839.539293] env[63088]: DEBUG nova.compute.manager [req-8938092b-c048-4db6-8ccb-d3b974563c97 req-63d4dffb-3076-4c94-9671-d98d360b4e08 service nova] [instance: 00925f94-dbf9-453c-a124-d8434679aedf] Received event network-changed-e2779ab6-aee9-4245-aab0-9d598a612844 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 839.539558] env[63088]: DEBUG nova.compute.manager [req-8938092b-c048-4db6-8ccb-d3b974563c97 req-63d4dffb-3076-4c94-9671-d98d360b4e08 service nova] [instance: 00925f94-dbf9-453c-a124-d8434679aedf] Refreshing instance network info cache due to event network-changed-e2779ab6-aee9-4245-aab0-9d598a612844. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 839.539930] env[63088]: DEBUG oslo_concurrency.lockutils [req-8938092b-c048-4db6-8ccb-d3b974563c97 req-63d4dffb-3076-4c94-9671-d98d360b4e08 service nova] Acquiring lock "refresh_cache-00925f94-dbf9-453c-a124-d8434679aedf" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 839.560469] env[63088]: DEBUG nova.compute.manager [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 7f341d80-450a-47b8-b26d-15f2ce3e378c] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 839.589783] env[63088]: DEBUG nova.virt.hardware [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 839.590052] env[63088]: DEBUG nova.virt.hardware [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 839.590245] env[63088]: DEBUG nova.virt.hardware [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 839.590438] env[63088]: DEBUG nova.virt.hardware [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 839.590582] env[63088]: DEBUG nova.virt.hardware [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 839.590793] env[63088]: DEBUG nova.virt.hardware [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 839.590954] env[63088]: DEBUG nova.virt.hardware [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 839.591133] env[63088]: DEBUG nova.virt.hardware [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 839.591304] env[63088]: DEBUG nova.virt.hardware [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 839.591466] env[63088]: DEBUG nova.virt.hardware [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 839.591667] env[63088]: DEBUG nova.virt.hardware [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 839.592576] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc6a1f3d-52f6-4fa5-9282-7d3fb9bd8ab9 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.601753] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c42ec09-b4dd-4495-880c-5480d7b8119f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.698060] env[63088]: DEBUG oslo_concurrency.lockutils [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Releasing lock "refresh_cache-00925f94-dbf9-453c-a124-d8434679aedf" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 839.698407] env[63088]: DEBUG nova.compute.manager [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 00925f94-dbf9-453c-a124-d8434679aedf] Instance network_info: |[{"id": "e2779ab6-aee9-4245-aab0-9d598a612844", "address": "fa:16:3e:da:19:08", "network": {"id": "79155637-c844-4012-a77a-d04f9ec9ab18", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-2144942540-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e27be174c0b3400193f9ae64a242da1a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape2779ab6-ae", "ovs_interfaceid": "e2779ab6-aee9-4245-aab0-9d598a612844", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 839.698707] env[63088]: DEBUG oslo_concurrency.lockutils [req-8938092b-c048-4db6-8ccb-d3b974563c97 req-63d4dffb-3076-4c94-9671-d98d360b4e08 service nova] Acquired lock "refresh_cache-00925f94-dbf9-453c-a124-d8434679aedf" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 839.698888] env[63088]: DEBUG nova.network.neutron [req-8938092b-c048-4db6-8ccb-d3b974563c97 req-63d4dffb-3076-4c94-9671-d98d360b4e08 service nova] [instance: 00925f94-dbf9-453c-a124-d8434679aedf] Refreshing network info cache for port e2779ab6-aee9-4245-aab0-9d598a612844 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 839.700143] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 00925f94-dbf9-453c-a124-d8434679aedf] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:da:19:08', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '93c5b7ce-4c84-40bc-884c-b2453e0eee69', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e2779ab6-aee9-4245-aab0-9d598a612844', 'vif_model': 'vmxnet3'}] {{(pid=63088) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 839.707595] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Creating folder: Project (e27be174c0b3400193f9ae64a242da1a). Parent ref: group-v275816. {{(pid=63088) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 839.712638] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d46c9cbb-8d7f-4980-82cb-b5d251f03d92 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.724210] env[63088]: INFO nova.virt.vmwareapi.vm_util [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Created folder: Project (e27be174c0b3400193f9ae64a242da1a) in parent group-v275816. [ 839.724411] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Creating folder: Instances. Parent ref: group-v275868. {{(pid=63088) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 839.724650] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c0c8feb1-3b2a-4252-a4c9-fa5ce29fc85b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.728923] env[63088]: INFO nova.compute.manager [-] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] Took 1.26 seconds to deallocate network for instance. [ 839.735541] env[63088]: INFO nova.virt.vmwareapi.vm_util [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Created folder: Instances in parent group-v275868. [ 839.735642] env[63088]: DEBUG oslo.service.loopingcall [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 839.736656] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 00925f94-dbf9-453c-a124-d8434679aedf] Creating VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 839.736656] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d960995f-3c9e-4081-af67-655fa32ecdd1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.758048] env[63088]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 839.758048] env[63088]: value = "task-1284959" [ 839.758048] env[63088]: _type = "Task" [ 839.758048] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.768728] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1284959, 'name': CreateVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.794236] env[63088]: DEBUG oslo_vmware.api [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Task: {'id': task-1284955, 'name': Rename_Task, 'duration_secs': 0.17295} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.798148] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] [instance: b159649b-ebd5-4c7d-9074-dc0e6395a947] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 839.798451] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-43f53207-0112-4798-8366-b5fefa0541ed {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.805722] env[63088]: DEBUG oslo_vmware.api [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Task: {'id': task-1284956, 'name': ReconfigVM_Task, 'duration_secs': 0.299429} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.809833] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] [instance: 2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68] Reconfigured VM instance instance-0000003e to attach disk [datastore2] 2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68/2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 839.810638] env[63088]: DEBUG oslo_vmware.api [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Waiting for the task: (returnval){ [ 839.810638] env[63088]: value = "task-1284960" [ 839.810638] env[63088]: _type = "Task" [ 839.810638] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.810778] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-67b67935-d333-4d68-ba8f-3c9bc6f970bf {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.821799] env[63088]: DEBUG oslo_vmware.api [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Task: {'id': task-1284954, 'name': ReconfigVM_Task, 'duration_secs': 0.621347} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.822457] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] [instance: 1d421bf0-925d-408c-a612-b659da92799f] Reconfigured VM instance instance-0000003d to attach disk [datastore1] 1d421bf0-925d-408c-a612-b659da92799f/1d421bf0-925d-408c-a612-b659da92799f.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 839.823230] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ba0ca1bd-7d05-486c-a105-b7073b99698d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.829401] env[63088]: DEBUG oslo_vmware.api [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Task: {'id': task-1284960, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.829697] env[63088]: DEBUG oslo_vmware.api [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Waiting for the task: (returnval){ [ 839.829697] env[63088]: value = "task-1284961" [ 839.829697] env[63088]: _type = "Task" [ 839.829697] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.834711] env[63088]: DEBUG oslo_vmware.api [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Waiting for the task: (returnval){ [ 839.834711] env[63088]: value = "task-1284962" [ 839.834711] env[63088]: _type = "Task" [ 839.834711] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.841860] env[63088]: DEBUG oslo_vmware.api [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Task: {'id': task-1284961, 'name': Rename_Task} progress is 10%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.846975] env[63088]: DEBUG oslo_vmware.api [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Task: {'id': task-1284962, 'name': Rename_Task} progress is 6%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.019232] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.485s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.019959] env[63088]: DEBUG nova.compute.manager [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 8b564e29-96cf-4abf-963d-142b413fb464] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 840.023419] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.895s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.025221] env[63088]: INFO nova.compute.claims [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 840.070926] env[63088]: DEBUG nova.network.neutron [req-8938092b-c048-4db6-8ccb-d3b974563c97 req-63d4dffb-3076-4c94-9671-d98d360b4e08 service nova] [instance: 00925f94-dbf9-453c-a124-d8434679aedf] Updated VIF entry in instance network info cache for port e2779ab6-aee9-4245-aab0-9d598a612844. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 840.071298] env[63088]: DEBUG nova.network.neutron [req-8938092b-c048-4db6-8ccb-d3b974563c97 req-63d4dffb-3076-4c94-9671-d98d360b4e08 service nova] [instance: 00925f94-dbf9-453c-a124-d8434679aedf] Updating instance_info_cache with network_info: [{"id": "e2779ab6-aee9-4245-aab0-9d598a612844", "address": "fa:16:3e:da:19:08", "network": {"id": "79155637-c844-4012-a77a-d04f9ec9ab18", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-2144942540-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e27be174c0b3400193f9ae64a242da1a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape2779ab6-ae", "ovs_interfaceid": "e2779ab6-aee9-4245-aab0-9d598a612844", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 840.237098] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0d443006-c001-488d-9c7f-f5cbe25ea71a tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 840.268511] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1284959, 'name': CreateVM_Task, 'duration_secs': 0.363324} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.268700] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 00925f94-dbf9-453c-a124-d8434679aedf] Created VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 840.269373] env[63088]: DEBUG oslo_concurrency.lockutils [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 840.269533] env[63088]: DEBUG oslo_concurrency.lockutils [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 840.269881] env[63088]: DEBUG oslo_concurrency.lockutils [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 840.270142] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e0a64a51-cbe5-4be1-a68e-306e24aeb9e5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.274482] env[63088]: DEBUG oslo_vmware.api [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Waiting for the task: (returnval){ [ 840.274482] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]5286d342-8e7c-b61f-3693-a298f1d41b95" [ 840.274482] env[63088]: _type = "Task" [ 840.274482] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.281708] env[63088]: DEBUG oslo_vmware.api [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]5286d342-8e7c-b61f-3693-a298f1d41b95, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.321048] env[63088]: DEBUG oslo_vmware.api [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Task: {'id': task-1284960, 'name': PowerOnVM_Task} progress is 87%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.338332] env[63088]: DEBUG oslo_vmware.api [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Task: {'id': task-1284961, 'name': Rename_Task, 'duration_secs': 0.145274} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.341142] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] [instance: 2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 840.341670] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4409468e-6912-4afe-b03c-d15720c9ecba {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.348012] env[63088]: DEBUG oslo_vmware.api [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Task: {'id': task-1284962, 'name': Rename_Task, 'duration_secs': 0.182048} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.349089] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] [instance: 1d421bf0-925d-408c-a612-b659da92799f] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 840.349414] env[63088]: DEBUG oslo_vmware.api [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Waiting for the task: (returnval){ [ 840.349414] env[63088]: value = "task-1284963" [ 840.349414] env[63088]: _type = "Task" [ 840.349414] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.349700] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8deaacd0-9938-495a-940c-28fabdb71e40 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.358492] env[63088]: DEBUG oslo_vmware.api [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Task: {'id': task-1284963, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.359514] env[63088]: DEBUG oslo_vmware.api [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Waiting for the task: (returnval){ [ 840.359514] env[63088]: value = "task-1284964" [ 840.359514] env[63088]: _type = "Task" [ 840.359514] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.366548] env[63088]: DEBUG oslo_vmware.api [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Task: {'id': task-1284964, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.533048] env[63088]: DEBUG nova.compute.utils [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 840.534450] env[63088]: DEBUG nova.compute.manager [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 8b564e29-96cf-4abf-963d-142b413fb464] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 840.534615] env[63088]: DEBUG nova.network.neutron [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 8b564e29-96cf-4abf-963d-142b413fb464] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 840.574277] env[63088]: DEBUG oslo_concurrency.lockutils [req-8938092b-c048-4db6-8ccb-d3b974563c97 req-63d4dffb-3076-4c94-9671-d98d360b4e08 service nova] Releasing lock "refresh_cache-00925f94-dbf9-453c-a124-d8434679aedf" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 840.574591] env[63088]: DEBUG nova.compute.manager [req-8938092b-c048-4db6-8ccb-d3b974563c97 req-63d4dffb-3076-4c94-9671-d98d360b4e08 service nova] [instance: 34ce411f-40c8-446e-b685-cd3ce07663dd] Received event network-changed-44d58f1e-f90f-4aea-b8fd-5f9c7ff195cb {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 840.574784] env[63088]: DEBUG nova.compute.manager [req-8938092b-c048-4db6-8ccb-d3b974563c97 req-63d4dffb-3076-4c94-9671-d98d360b4e08 service nova] [instance: 34ce411f-40c8-446e-b685-cd3ce07663dd] Refreshing instance network info cache due to event network-changed-44d58f1e-f90f-4aea-b8fd-5f9c7ff195cb. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 840.575076] env[63088]: DEBUG oslo_concurrency.lockutils [req-8938092b-c048-4db6-8ccb-d3b974563c97 req-63d4dffb-3076-4c94-9671-d98d360b4e08 service nova] Acquiring lock "refresh_cache-34ce411f-40c8-446e-b685-cd3ce07663dd" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 840.575279] env[63088]: DEBUG oslo_concurrency.lockutils [req-8938092b-c048-4db6-8ccb-d3b974563c97 req-63d4dffb-3076-4c94-9671-d98d360b4e08 service nova] Acquired lock "refresh_cache-34ce411f-40c8-446e-b685-cd3ce07663dd" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 840.575475] env[63088]: DEBUG nova.network.neutron [req-8938092b-c048-4db6-8ccb-d3b974563c97 req-63d4dffb-3076-4c94-9671-d98d360b4e08 service nova] [instance: 34ce411f-40c8-446e-b685-cd3ce07663dd] Refreshing network info cache for port 44d58f1e-f90f-4aea-b8fd-5f9c7ff195cb {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 840.622955] env[63088]: DEBUG nova.policy [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a6f91fc8922c42f985f6f6338dab0708', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7bff1382b9694df08133c88a5fe783a5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 840.786460] env[63088]: DEBUG oslo_vmware.api [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]5286d342-8e7c-b61f-3693-a298f1d41b95, 'name': SearchDatastore_Task, 'duration_secs': 0.036459} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.787043] env[63088]: DEBUG oslo_concurrency.lockutils [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 840.787043] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 00925f94-dbf9-453c-a124-d8434679aedf] Processing image 1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 840.787278] env[63088]: DEBUG oslo_concurrency.lockutils [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 840.787430] env[63088]: DEBUG oslo_concurrency.lockutils [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 840.787660] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 840.787962] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0a2e6813-8fb5-41e1-993e-e8468b4982fe {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.798317] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 840.798539] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 840.799471] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5ab7e4d4-48c8-458e-aa3d-8acc6bcbde1c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.807846] env[63088]: DEBUG oslo_vmware.api [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Waiting for the task: (returnval){ [ 840.807846] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52199c84-7a1e-33a6-5997-8a6509620314" [ 840.807846] env[63088]: _type = "Task" [ 840.807846] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.817022] env[63088]: DEBUG oslo_vmware.api [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52199c84-7a1e-33a6-5997-8a6509620314, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.825474] env[63088]: DEBUG oslo_vmware.api [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Task: {'id': task-1284960, 'name': PowerOnVM_Task, 'duration_secs': 0.805523} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.827020] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] [instance: b159649b-ebd5-4c7d-9074-dc0e6395a947] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 840.827020] env[63088]: DEBUG nova.compute.manager [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] [instance: b159649b-ebd5-4c7d-9074-dc0e6395a947] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 840.827020] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11e74fa2-434b-4310-a005-8fc021638d76 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.866198] env[63088]: DEBUG oslo_vmware.api [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Task: {'id': task-1284963, 'name': PowerOnVM_Task, 'duration_secs': 0.457032} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.867232] env[63088]: DEBUG nova.network.neutron [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 7f341d80-450a-47b8-b26d-15f2ce3e378c] Successfully updated port: 711f2731-680e-4328-97c8-0a2d1bbf79eb {{(pid=63088) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 840.875926] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] [instance: 2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 840.876194] env[63088]: INFO nova.compute.manager [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] [instance: 2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68] Took 8.07 seconds to spawn the instance on the hypervisor. [ 840.876731] env[63088]: DEBUG nova.compute.manager [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] [instance: 2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 840.878260] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e13c1c5-94e0-4ff8-bd2c-4e3c993c4761 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.889835] env[63088]: DEBUG oslo_vmware.api [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Task: {'id': task-1284964, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.038228] env[63088]: DEBUG nova.compute.manager [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 8b564e29-96cf-4abf-963d-142b413fb464] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 841.110194] env[63088]: DEBUG nova.network.neutron [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 8b564e29-96cf-4abf-963d-142b413fb464] Successfully created port: 7a0cbd40-cac8-4bb6-944c-486989457770 {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 841.325204] env[63088]: DEBUG oslo_vmware.api [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52199c84-7a1e-33a6-5997-8a6509620314, 'name': SearchDatastore_Task, 'duration_secs': 0.013991} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.325204] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cfbf5296-a285-4f76-9281-11ff4e94fa2e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.330516] env[63088]: DEBUG oslo_vmware.api [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Waiting for the task: (returnval){ [ 841.330516] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52a4bea9-8570-ae0e-cc71-400e71cddc9a" [ 841.330516] env[63088]: _type = "Task" [ 841.330516] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.342895] env[63088]: DEBUG oslo_vmware.api [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52a4bea9-8570-ae0e-cc71-400e71cddc9a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.350383] env[63088]: DEBUG oslo_concurrency.lockutils [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 841.373534] env[63088]: DEBUG oslo_vmware.api [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Task: {'id': task-1284964, 'name': PowerOnVM_Task, 'duration_secs': 0.629065} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.376396] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] [instance: 1d421bf0-925d-408c-a612-b659da92799f] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 841.376608] env[63088]: INFO nova.compute.manager [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] [instance: 1d421bf0-925d-408c-a612-b659da92799f] Took 10.97 seconds to spawn the instance on the hypervisor. [ 841.376785] env[63088]: DEBUG nova.compute.manager [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] [instance: 1d421bf0-925d-408c-a612-b659da92799f] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 841.378162] env[63088]: DEBUG oslo_concurrency.lockutils [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Acquiring lock "refresh_cache-7f341d80-450a-47b8-b26d-15f2ce3e378c" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 841.378162] env[63088]: DEBUG oslo_concurrency.lockutils [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Acquired lock "refresh_cache-7f341d80-450a-47b8-b26d-15f2ce3e378c" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 841.378162] env[63088]: DEBUG nova.network.neutron [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 7f341d80-450a-47b8-b26d-15f2ce3e378c] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 841.382457] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36242dd4-4c13-4dac-b5f8-6ee0f0b95d23 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.411167] env[63088]: INFO nova.compute.manager [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] [instance: 2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68] Took 21.91 seconds to build instance. [ 841.459807] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ca2df92-3c77-4a22-a42d-2b29106ab4ba {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.468320] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcf98b73-a666-44bf-a6d8-82681ffbc65d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.509392] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbc2054a-5d79-4f41-9ab9-f4a1c57ef4e9 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.515465] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e27a8b01-946b-4c58-b3f1-573efcc59b96 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.529796] env[63088]: DEBUG nova.compute.provider_tree [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 841.605775] env[63088]: DEBUG nova.network.neutron [req-8938092b-c048-4db6-8ccb-d3b974563c97 req-63d4dffb-3076-4c94-9671-d98d360b4e08 service nova] [instance: 34ce411f-40c8-446e-b685-cd3ce07663dd] Updated VIF entry in instance network info cache for port 44d58f1e-f90f-4aea-b8fd-5f9c7ff195cb. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 841.606183] env[63088]: DEBUG nova.network.neutron [req-8938092b-c048-4db6-8ccb-d3b974563c97 req-63d4dffb-3076-4c94-9671-d98d360b4e08 service nova] [instance: 34ce411f-40c8-446e-b685-cd3ce07663dd] Updating instance_info_cache with network_info: [{"id": "44d58f1e-f90f-4aea-b8fd-5f9c7ff195cb", "address": "fa:16:3e:bb:b4:13", "network": {"id": "88a981da-0420-42f7-bc5c-0bfb53e39272", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-1197986229-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.225", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a5f62816dd0243da9be2073086b136d8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92fe29b3-0907-453d-aabb-5559c4bd7c0f", "external-id": "nsx-vlan-transportzone-482", "segmentation_id": 482, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap44d58f1e-f9", "ovs_interfaceid": "44d58f1e-f90f-4aea-b8fd-5f9c7ff195cb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 841.840764] env[63088]: DEBUG oslo_vmware.api [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52a4bea9-8570-ae0e-cc71-400e71cddc9a, 'name': SearchDatastore_Task, 'duration_secs': 0.012128} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.842135] env[63088]: DEBUG oslo_concurrency.lockutils [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 841.842422] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore2] 00925f94-dbf9-453c-a124-d8434679aedf/00925f94-dbf9-453c-a124-d8434679aedf.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 841.843890] env[63088]: DEBUG nova.compute.manager [req-184a9df8-439c-4d79-a830-d41fe039dad1 req-70d554eb-689c-4844-b672-6cb1a70737a2 service nova] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Received event network-changed-0c182796-7fb6-40ab-ba27-a43f0700dee1 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 841.846407] env[63088]: DEBUG nova.compute.manager [req-184a9df8-439c-4d79-a830-d41fe039dad1 req-70d554eb-689c-4844-b672-6cb1a70737a2 service nova] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Refreshing instance network info cache due to event network-changed-0c182796-7fb6-40ab-ba27-a43f0700dee1. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 841.846407] env[63088]: DEBUG oslo_concurrency.lockutils [req-184a9df8-439c-4d79-a830-d41fe039dad1 req-70d554eb-689c-4844-b672-6cb1a70737a2 service nova] Acquiring lock "refresh_cache-07fd3e6f-e0af-467c-9039-238be5c58d25" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 841.846407] env[63088]: DEBUG oslo_concurrency.lockutils [req-184a9df8-439c-4d79-a830-d41fe039dad1 req-70d554eb-689c-4844-b672-6cb1a70737a2 service nova] Acquired lock "refresh_cache-07fd3e6f-e0af-467c-9039-238be5c58d25" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 841.846407] env[63088]: DEBUG nova.network.neutron [req-184a9df8-439c-4d79-a830-d41fe039dad1 req-70d554eb-689c-4844-b672-6cb1a70737a2 service nova] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Refreshing network info cache for port 0c182796-7fb6-40ab-ba27-a43f0700dee1 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 841.846407] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cef4b607-18fe-48b7-af54-fbeb72c6c4be {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.853720] env[63088]: DEBUG oslo_vmware.api [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Waiting for the task: (returnval){ [ 841.853720] env[63088]: value = "task-1284965" [ 841.853720] env[63088]: _type = "Task" [ 841.853720] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.871830] env[63088]: DEBUG oslo_vmware.api [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': task-1284965, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.912402] env[63088]: DEBUG oslo_concurrency.lockutils [None req-32689d0f-d9d5-4bef-bc8b-b618cc122413 tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Lock "2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 56.107s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 841.919064] env[63088]: INFO nova.compute.manager [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] [instance: 1d421bf0-925d-408c-a612-b659da92799f] Took 25.09 seconds to build instance. [ 841.937742] env[63088]: DEBUG nova.network.neutron [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 7f341d80-450a-47b8-b26d-15f2ce3e378c] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 842.012228] env[63088]: DEBUG oslo_concurrency.lockutils [None req-8102b1ea-81e5-4196-b4f7-a4fd361924a7 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Acquiring lock "b159649b-ebd5-4c7d-9074-dc0e6395a947" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.012492] env[63088]: DEBUG oslo_concurrency.lockutils [None req-8102b1ea-81e5-4196-b4f7-a4fd361924a7 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Lock "b159649b-ebd5-4c7d-9074-dc0e6395a947" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 842.012693] env[63088]: DEBUG oslo_concurrency.lockutils [None req-8102b1ea-81e5-4196-b4f7-a4fd361924a7 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Acquiring lock "b159649b-ebd5-4c7d-9074-dc0e6395a947-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.012864] env[63088]: DEBUG oslo_concurrency.lockutils [None req-8102b1ea-81e5-4196-b4f7-a4fd361924a7 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Lock "b159649b-ebd5-4c7d-9074-dc0e6395a947-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 842.013030] env[63088]: DEBUG oslo_concurrency.lockutils [None req-8102b1ea-81e5-4196-b4f7-a4fd361924a7 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Lock "b159649b-ebd5-4c7d-9074-dc0e6395a947-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 842.015791] env[63088]: INFO nova.compute.manager [None req-8102b1ea-81e5-4196-b4f7-a4fd361924a7 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] [instance: b159649b-ebd5-4c7d-9074-dc0e6395a947] Terminating instance [ 842.019342] env[63088]: DEBUG oslo_concurrency.lockutils [None req-8102b1ea-81e5-4196-b4f7-a4fd361924a7 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Acquiring lock "refresh_cache-b159649b-ebd5-4c7d-9074-dc0e6395a947" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 842.019492] env[63088]: DEBUG oslo_concurrency.lockutils [None req-8102b1ea-81e5-4196-b4f7-a4fd361924a7 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Acquired lock "refresh_cache-b159649b-ebd5-4c7d-9074-dc0e6395a947" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 842.019689] env[63088]: DEBUG nova.network.neutron [None req-8102b1ea-81e5-4196-b4f7-a4fd361924a7 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] [instance: b159649b-ebd5-4c7d-9074-dc0e6395a947] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 842.032724] env[63088]: DEBUG nova.scheduler.client.report [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 842.049463] env[63088]: DEBUG nova.compute.manager [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 8b564e29-96cf-4abf-963d-142b413fb464] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 842.080534] env[63088]: DEBUG nova.virt.hardware [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 842.081036] env[63088]: DEBUG nova.virt.hardware [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 842.081036] env[63088]: DEBUG nova.virt.hardware [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 842.081172] env[63088]: DEBUG nova.virt.hardware [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 842.081370] env[63088]: DEBUG nova.virt.hardware [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 842.081426] env[63088]: DEBUG nova.virt.hardware [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 842.081639] env[63088]: DEBUG nova.virt.hardware [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 842.081819] env[63088]: DEBUG nova.virt.hardware [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 842.082007] env[63088]: DEBUG nova.virt.hardware [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 842.082642] env[63088]: DEBUG nova.virt.hardware [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 842.082832] env[63088]: DEBUG nova.virt.hardware [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 842.083693] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e215bffb-67e0-4a97-a08b-4eb0f8f434f3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.094301] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92939ede-fe6f-48d7-a215-bc385ea10c58 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.109336] env[63088]: DEBUG oslo_concurrency.lockutils [req-8938092b-c048-4db6-8ccb-d3b974563c97 req-63d4dffb-3076-4c94-9671-d98d360b4e08 service nova] Releasing lock "refresh_cache-34ce411f-40c8-446e-b685-cd3ce07663dd" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 842.132171] env[63088]: DEBUG nova.network.neutron [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 7f341d80-450a-47b8-b26d-15f2ce3e378c] Updating instance_info_cache with network_info: [{"id": "711f2731-680e-4328-97c8-0a2d1bbf79eb", "address": "fa:16:3e:46:39:13", "network": {"id": "79155637-c844-4012-a77a-d04f9ec9ab18", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-2144942540-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e27be174c0b3400193f9ae64a242da1a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap711f2731-68", "ovs_interfaceid": "711f2731-680e-4328-97c8-0a2d1bbf79eb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 842.182572] env[63088]: DEBUG nova.compute.manager [req-1fffb8c2-a9b0-4f4b-affe-174c62ff656f req-2bbba38c-19b2-4181-9690-aa8db145e658 service nova] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Received event network-changed-0c182796-7fb6-40ab-ba27-a43f0700dee1 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 842.182572] env[63088]: DEBUG nova.compute.manager [req-1fffb8c2-a9b0-4f4b-affe-174c62ff656f req-2bbba38c-19b2-4181-9690-aa8db145e658 service nova] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Refreshing instance network info cache due to event network-changed-0c182796-7fb6-40ab-ba27-a43f0700dee1. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 842.182852] env[63088]: DEBUG oslo_concurrency.lockutils [req-1fffb8c2-a9b0-4f4b-affe-174c62ff656f req-2bbba38c-19b2-4181-9690-aa8db145e658 service nova] Acquiring lock "refresh_cache-07fd3e6f-e0af-467c-9039-238be5c58d25" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 842.365282] env[63088]: DEBUG oslo_vmware.api [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': task-1284965, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.416912] env[63088]: DEBUG nova.compute.manager [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 842.421758] env[63088]: DEBUG oslo_concurrency.lockutils [None req-13ba7be9-4959-4ea2-8b59-98699a1d4458 tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Lock "1d421bf0-925d-408c-a612-b659da92799f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 68.004s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 842.538460] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.515s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 842.539023] env[63088]: DEBUG nova.compute.manager [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 842.544932] env[63088]: DEBUG nova.network.neutron [None req-8102b1ea-81e5-4196-b4f7-a4fd361924a7 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] [instance: b159649b-ebd5-4c7d-9074-dc0e6395a947] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 842.547511] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.464s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 842.549339] env[63088]: INFO nova.compute.claims [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 842.609285] env[63088]: DEBUG nova.network.neutron [None req-8102b1ea-81e5-4196-b4f7-a4fd361924a7 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] [instance: b159649b-ebd5-4c7d-9074-dc0e6395a947] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 842.635052] env[63088]: DEBUG oslo_concurrency.lockutils [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Releasing lock "refresh_cache-7f341d80-450a-47b8-b26d-15f2ce3e378c" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 842.635443] env[63088]: DEBUG nova.compute.manager [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 7f341d80-450a-47b8-b26d-15f2ce3e378c] Instance network_info: |[{"id": "711f2731-680e-4328-97c8-0a2d1bbf79eb", "address": "fa:16:3e:46:39:13", "network": {"id": "79155637-c844-4012-a77a-d04f9ec9ab18", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-2144942540-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e27be174c0b3400193f9ae64a242da1a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap711f2731-68", "ovs_interfaceid": "711f2731-680e-4328-97c8-0a2d1bbf79eb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 842.636376] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 7f341d80-450a-47b8-b26d-15f2ce3e378c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:46:39:13', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '93c5b7ce-4c84-40bc-884c-b2453e0eee69', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '711f2731-680e-4328-97c8-0a2d1bbf79eb', 'vif_model': 'vmxnet3'}] {{(pid=63088) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 842.644632] env[63088]: DEBUG oslo.service.loopingcall [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 842.645524] env[63088]: DEBUG nova.network.neutron [req-184a9df8-439c-4d79-a830-d41fe039dad1 req-70d554eb-689c-4844-b672-6cb1a70737a2 service nova] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Updated VIF entry in instance network info cache for port 0c182796-7fb6-40ab-ba27-a43f0700dee1. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 842.645854] env[63088]: DEBUG nova.network.neutron [req-184a9df8-439c-4d79-a830-d41fe039dad1 req-70d554eb-689c-4844-b672-6cb1a70737a2 service nova] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Updating instance_info_cache with network_info: [{"id": "0c182796-7fb6-40ab-ba27-a43f0700dee1", "address": "fa:16:3e:63:25:2a", "network": {"id": "de116097-3737-4dcf-972c-88c49fa8995d", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-580707267-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.168", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "937eba1a40c84534b279d5798e234625", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3a80436-f7a9-431a-acec-aca3d76e3f9b", "external-id": "cl2-zone-339", "segmentation_id": 339, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c182796-7f", "ovs_interfaceid": "0c182796-7fb6-40ab-ba27-a43f0700dee1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 842.647249] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7f341d80-450a-47b8-b26d-15f2ce3e378c] Creating VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 842.647249] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-69aa18b8-01ee-4fe3-a21e-3e0b43811f1e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.664123] env[63088]: DEBUG oslo_concurrency.lockutils [req-184a9df8-439c-4d79-a830-d41fe039dad1 req-70d554eb-689c-4844-b672-6cb1a70737a2 service nova] Releasing lock "refresh_cache-07fd3e6f-e0af-467c-9039-238be5c58d25" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 842.664370] env[63088]: DEBUG nova.compute.manager [req-184a9df8-439c-4d79-a830-d41fe039dad1 req-70d554eb-689c-4844-b672-6cb1a70737a2 service nova] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Received event network-changed-0c182796-7fb6-40ab-ba27-a43f0700dee1 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 842.664534] env[63088]: DEBUG nova.compute.manager [req-184a9df8-439c-4d79-a830-d41fe039dad1 req-70d554eb-689c-4844-b672-6cb1a70737a2 service nova] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Refreshing instance network info cache due to event network-changed-0c182796-7fb6-40ab-ba27-a43f0700dee1. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 842.664716] env[63088]: DEBUG oslo_concurrency.lockutils [req-184a9df8-439c-4d79-a830-d41fe039dad1 req-70d554eb-689c-4844-b672-6cb1a70737a2 service nova] Acquiring lock "refresh_cache-07fd3e6f-e0af-467c-9039-238be5c58d25" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 842.664854] env[63088]: DEBUG oslo_concurrency.lockutils [req-184a9df8-439c-4d79-a830-d41fe039dad1 req-70d554eb-689c-4844-b672-6cb1a70737a2 service nova] Acquired lock "refresh_cache-07fd3e6f-e0af-467c-9039-238be5c58d25" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 842.665027] env[63088]: DEBUG nova.network.neutron [req-184a9df8-439c-4d79-a830-d41fe039dad1 req-70d554eb-689c-4844-b672-6cb1a70737a2 service nova] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Refreshing network info cache for port 0c182796-7fb6-40ab-ba27-a43f0700dee1 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 842.671331] env[63088]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 842.671331] env[63088]: value = "task-1284966" [ 842.671331] env[63088]: _type = "Task" [ 842.671331] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.679701] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1284966, 'name': CreateVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.864832] env[63088]: DEBUG oslo_vmware.api [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': task-1284965, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.777923} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.865234] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore2] 00925f94-dbf9-453c-a124-d8434679aedf/00925f94-dbf9-453c-a124-d8434679aedf.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 842.865528] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 00925f94-dbf9-453c-a124-d8434679aedf] Extending root virtual disk to 1048576 {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 842.865858] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ef8f7435-32e1-4fbc-b539-54671b02cd3b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.872775] env[63088]: DEBUG oslo_vmware.api [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Waiting for the task: (returnval){ [ 842.872775] env[63088]: value = "task-1284967" [ 842.872775] env[63088]: _type = "Task" [ 842.872775] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.881566] env[63088]: DEBUG oslo_vmware.api [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': task-1284967, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.943184] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.950788] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b548d275-1b16-4d7d-aec5-f1421ffffdea tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Acquiring lock "2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.951381] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b548d275-1b16-4d7d-aec5-f1421ffffdea tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Lock "2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 842.951381] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b548d275-1b16-4d7d-aec5-f1421ffffdea tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Acquiring lock "2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.951505] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b548d275-1b16-4d7d-aec5-f1421ffffdea tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Lock "2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 842.951583] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b548d275-1b16-4d7d-aec5-f1421ffffdea tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Lock "2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 842.953682] env[63088]: INFO nova.compute.manager [None req-b548d275-1b16-4d7d-aec5-f1421ffffdea tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] [instance: 2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68] Terminating instance [ 842.955519] env[63088]: DEBUG nova.compute.manager [None req-b548d275-1b16-4d7d-aec5-f1421ffffdea tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] [instance: 2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 842.955588] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-b548d275-1b16-4d7d-aec5-f1421ffffdea tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] [instance: 2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 842.956395] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bba742a-e694-4ef2-8366-2cddbf163d40 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.963873] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-b548d275-1b16-4d7d-aec5-f1421ffffdea tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] [instance: 2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 842.964125] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-eb49fdb3-f64c-4498-9716-7edbe2f719be {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.966187] env[63088]: DEBUG nova.network.neutron [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 8b564e29-96cf-4abf-963d-142b413fb464] Successfully updated port: 7a0cbd40-cac8-4bb6-944c-486989457770 {{(pid=63088) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 842.971607] env[63088]: DEBUG oslo_vmware.api [None req-b548d275-1b16-4d7d-aec5-f1421ffffdea tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Waiting for the task: (returnval){ [ 842.971607] env[63088]: value = "task-1284968" [ 842.971607] env[63088]: _type = "Task" [ 842.971607] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.979866] env[63088]: DEBUG oslo_vmware.api [None req-b548d275-1b16-4d7d-aec5-f1421ffffdea tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Task: {'id': task-1284968, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.057342] env[63088]: DEBUG nova.compute.utils [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 843.059109] env[63088]: DEBUG nova.compute.manager [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 843.059346] env[63088]: DEBUG nova.network.neutron [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 843.103386] env[63088]: DEBUG nova.policy [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '787d7bd9f8e7490c9f24ff0e159bc98f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'dace8b5181b84623b08f903d12dfd31e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 843.112652] env[63088]: DEBUG oslo_concurrency.lockutils [None req-8102b1ea-81e5-4196-b4f7-a4fd361924a7 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Releasing lock "refresh_cache-b159649b-ebd5-4c7d-9074-dc0e6395a947" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 843.112652] env[63088]: DEBUG nova.compute.manager [None req-8102b1ea-81e5-4196-b4f7-a4fd361924a7 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] [instance: b159649b-ebd5-4c7d-9074-dc0e6395a947] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 843.112834] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-8102b1ea-81e5-4196-b4f7-a4fd361924a7 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] [instance: b159649b-ebd5-4c7d-9074-dc0e6395a947] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 843.113693] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a16dcfd0-436f-4db7-be0c-9fbb6ed14ddc {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.121697] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-8102b1ea-81e5-4196-b4f7-a4fd361924a7 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] [instance: b159649b-ebd5-4c7d-9074-dc0e6395a947] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 843.121968] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c59e3fc9-a94e-4430-a465-b3caf50d67b6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.128403] env[63088]: DEBUG oslo_vmware.api [None req-8102b1ea-81e5-4196-b4f7-a4fd361924a7 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Waiting for the task: (returnval){ [ 843.128403] env[63088]: value = "task-1284969" [ 843.128403] env[63088]: _type = "Task" [ 843.128403] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.136511] env[63088]: DEBUG oslo_vmware.api [None req-8102b1ea-81e5-4196-b4f7-a4fd361924a7 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Task: {'id': task-1284969, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.181663] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1284966, 'name': CreateVM_Task, 'duration_secs': 0.437085} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.181872] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7f341d80-450a-47b8-b26d-15f2ce3e378c] Created VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 843.182629] env[63088]: DEBUG oslo_concurrency.lockutils [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 843.182843] env[63088]: DEBUG oslo_concurrency.lockutils [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.183225] env[63088]: DEBUG oslo_concurrency.lockutils [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 843.183527] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-072b4dc7-45a9-4271-910d-682b8305b12d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.188793] env[63088]: DEBUG oslo_vmware.api [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Waiting for the task: (returnval){ [ 843.188793] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]5295a17a-b667-1893-e1bb-aa191566ad04" [ 843.188793] env[63088]: _type = "Task" [ 843.188793] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.197582] env[63088]: DEBUG oslo_vmware.api [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]5295a17a-b667-1893-e1bb-aa191566ad04, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.384686] env[63088]: DEBUG oslo_vmware.api [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': task-1284967, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062721} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.384936] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 00925f94-dbf9-453c-a124-d8434679aedf] Extended root virtual disk {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 843.385949] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29ccfb8c-9532-4e2e-8a07-258f27341b63 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.418730] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 00925f94-dbf9-453c-a124-d8434679aedf] Reconfiguring VM instance instance-0000003f to attach disk [datastore2] 00925f94-dbf9-453c-a124-d8434679aedf/00925f94-dbf9-453c-a124-d8434679aedf.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 843.419065] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-914e0e36-d854-493e-be67-ce06b1b45660 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.445017] env[63088]: DEBUG oslo_vmware.api [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Waiting for the task: (returnval){ [ 843.445017] env[63088]: value = "task-1284970" [ 843.445017] env[63088]: _type = "Task" [ 843.445017] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.451305] env[63088]: DEBUG oslo_vmware.api [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': task-1284970, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.470660] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Acquiring lock "refresh_cache-8b564e29-96cf-4abf-963d-142b413fb464" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 843.470966] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Acquired lock "refresh_cache-8b564e29-96cf-4abf-963d-142b413fb464" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.475756] env[63088]: DEBUG nova.network.neutron [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 8b564e29-96cf-4abf-963d-142b413fb464] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 843.493297] env[63088]: DEBUG oslo_vmware.api [None req-b548d275-1b16-4d7d-aec5-f1421ffffdea tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Task: {'id': task-1284968, 'name': PowerOffVM_Task, 'duration_secs': 0.206524} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.494389] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-b548d275-1b16-4d7d-aec5-f1421ffffdea tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] [instance: 2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 843.494645] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-b548d275-1b16-4d7d-aec5-f1421ffffdea tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] [instance: 2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68] Unregistering the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 843.495058] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-023b21a2-7be1-464d-86d2-0300704dab97 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.557967] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-b548d275-1b16-4d7d-aec5-f1421ffffdea tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] [instance: 2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68] Unregistered the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 843.558241] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-b548d275-1b16-4d7d-aec5-f1421ffffdea tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] [instance: 2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68] Deleting contents of the VM from datastore datastore2 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 843.558452] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-b548d275-1b16-4d7d-aec5-f1421ffffdea tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Deleting the datastore file [datastore2] 2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68 {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 843.558736] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e3782128-31fc-4b8e-8f3e-919a81a4f4cd {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.564249] env[63088]: DEBUG nova.compute.manager [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 843.568384] env[63088]: DEBUG nova.network.neutron [req-184a9df8-439c-4d79-a830-d41fe039dad1 req-70d554eb-689c-4844-b672-6cb1a70737a2 service nova] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Updated VIF entry in instance network info cache for port 0c182796-7fb6-40ab-ba27-a43f0700dee1. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 843.568832] env[63088]: DEBUG nova.network.neutron [req-184a9df8-439c-4d79-a830-d41fe039dad1 req-70d554eb-689c-4844-b672-6cb1a70737a2 service nova] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Updating instance_info_cache with network_info: [{"id": "0c182796-7fb6-40ab-ba27-a43f0700dee1", "address": "fa:16:3e:63:25:2a", "network": {"id": "de116097-3737-4dcf-972c-88c49fa8995d", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-580707267-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "937eba1a40c84534b279d5798e234625", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3a80436-f7a9-431a-acec-aca3d76e3f9b", "external-id": "cl2-zone-339", "segmentation_id": 339, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c182796-7f", "ovs_interfaceid": "0c182796-7fb6-40ab-ba27-a43f0700dee1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 843.571360] env[63088]: DEBUG oslo_vmware.api [None req-b548d275-1b16-4d7d-aec5-f1421ffffdea tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Waiting for the task: (returnval){ [ 843.571360] env[63088]: value = "task-1284972" [ 843.571360] env[63088]: _type = "Task" [ 843.571360] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.579792] env[63088]: DEBUG oslo_vmware.api [None req-b548d275-1b16-4d7d-aec5-f1421ffffdea tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Task: {'id': task-1284972, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.645471] env[63088]: DEBUG oslo_vmware.api [None req-8102b1ea-81e5-4196-b4f7-a4fd361924a7 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Task: {'id': task-1284969, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.696370] env[63088]: DEBUG nova.network.neutron [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Successfully created port: fa4a91d1-df9a-4789-bc5c-a8b95457cd93 {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 843.706607] env[63088]: DEBUG oslo_vmware.api [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]5295a17a-b667-1893-e1bb-aa191566ad04, 'name': SearchDatastore_Task, 'duration_secs': 0.047265} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.706974] env[63088]: DEBUG oslo_concurrency.lockutils [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 843.707267] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 7f341d80-450a-47b8-b26d-15f2ce3e378c] Processing image 1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 843.707539] env[63088]: DEBUG oslo_concurrency.lockutils [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 843.707762] env[63088]: DEBUG oslo_concurrency.lockutils [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.707981] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 843.708348] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4319eb00-cc5b-4ba8-9758-dcc4e53b29ef {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.725924] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 843.726112] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 843.727429] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-86f72629-e570-4a68-b968-bc5506e69098 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.734194] env[63088]: DEBUG oslo_vmware.api [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Waiting for the task: (returnval){ [ 843.734194] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52cc0bbc-8cbc-04ce-b83a-7e9bfcf14b9a" [ 843.734194] env[63088]: _type = "Task" [ 843.734194] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.742050] env[63088]: DEBUG oslo_vmware.api [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52cc0bbc-8cbc-04ce-b83a-7e9bfcf14b9a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.861509] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18b3ea80-c70e-4139-8860-4ce277cfc573 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.871835] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eec4fbc2-009c-4108-b8df-c010fb575387 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.908495] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47100d57-8ea4-4c59-8b4d-3f360ccd348c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.916539] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fb884af-35e0-4b9d-b7fa-495d96f14da0 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.930545] env[63088]: DEBUG nova.compute.provider_tree [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 843.952959] env[63088]: DEBUG oslo_vmware.api [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': task-1284970, 'name': ReconfigVM_Task, 'duration_secs': 0.310733} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.953256] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 00925f94-dbf9-453c-a124-d8434679aedf] Reconfigured VM instance instance-0000003f to attach disk [datastore2] 00925f94-dbf9-453c-a124-d8434679aedf/00925f94-dbf9-453c-a124-d8434679aedf.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 843.953880] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b57a3f39-c197-45b8-ab90-bbb41fa4adc7 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.960664] env[63088]: DEBUG oslo_vmware.api [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Waiting for the task: (returnval){ [ 843.960664] env[63088]: value = "task-1284973" [ 843.960664] env[63088]: _type = "Task" [ 843.960664] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.970503] env[63088]: DEBUG oslo_vmware.api [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': task-1284973, 'name': Rename_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.060112] env[63088]: DEBUG nova.network.neutron [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 8b564e29-96cf-4abf-963d-142b413fb464] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 844.076174] env[63088]: DEBUG oslo_concurrency.lockutils [req-184a9df8-439c-4d79-a830-d41fe039dad1 req-70d554eb-689c-4844-b672-6cb1a70737a2 service nova] Releasing lock "refresh_cache-07fd3e6f-e0af-467c-9039-238be5c58d25" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.076174] env[63088]: DEBUG nova.compute.manager [req-184a9df8-439c-4d79-a830-d41fe039dad1 req-70d554eb-689c-4844-b672-6cb1a70737a2 service nova] [instance: 7f341d80-450a-47b8-b26d-15f2ce3e378c] Received event network-vif-plugged-711f2731-680e-4328-97c8-0a2d1bbf79eb {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 844.076174] env[63088]: DEBUG oslo_concurrency.lockutils [req-184a9df8-439c-4d79-a830-d41fe039dad1 req-70d554eb-689c-4844-b672-6cb1a70737a2 service nova] Acquiring lock "7f341d80-450a-47b8-b26d-15f2ce3e378c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.076174] env[63088]: DEBUG oslo_concurrency.lockutils [req-184a9df8-439c-4d79-a830-d41fe039dad1 req-70d554eb-689c-4844-b672-6cb1a70737a2 service nova] Lock "7f341d80-450a-47b8-b26d-15f2ce3e378c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.076174] env[63088]: DEBUG oslo_concurrency.lockutils [req-184a9df8-439c-4d79-a830-d41fe039dad1 req-70d554eb-689c-4844-b672-6cb1a70737a2 service nova] Lock "7f341d80-450a-47b8-b26d-15f2ce3e378c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.076523] env[63088]: DEBUG nova.compute.manager [req-184a9df8-439c-4d79-a830-d41fe039dad1 req-70d554eb-689c-4844-b672-6cb1a70737a2 service nova] [instance: 7f341d80-450a-47b8-b26d-15f2ce3e378c] No waiting events found dispatching network-vif-plugged-711f2731-680e-4328-97c8-0a2d1bbf79eb {{(pid=63088) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 844.077419] env[63088]: WARNING nova.compute.manager [req-184a9df8-439c-4d79-a830-d41fe039dad1 req-70d554eb-689c-4844-b672-6cb1a70737a2 service nova] [instance: 7f341d80-450a-47b8-b26d-15f2ce3e378c] Received unexpected event network-vif-plugged-711f2731-680e-4328-97c8-0a2d1bbf79eb for instance with vm_state building and task_state spawning. [ 844.077645] env[63088]: DEBUG nova.compute.manager [req-184a9df8-439c-4d79-a830-d41fe039dad1 req-70d554eb-689c-4844-b672-6cb1a70737a2 service nova] [instance: 7f341d80-450a-47b8-b26d-15f2ce3e378c] Received event network-changed-711f2731-680e-4328-97c8-0a2d1bbf79eb {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 844.077813] env[63088]: DEBUG nova.compute.manager [req-184a9df8-439c-4d79-a830-d41fe039dad1 req-70d554eb-689c-4844-b672-6cb1a70737a2 service nova] [instance: 7f341d80-450a-47b8-b26d-15f2ce3e378c] Refreshing instance network info cache due to event network-changed-711f2731-680e-4328-97c8-0a2d1bbf79eb. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 844.078021] env[63088]: DEBUG oslo_concurrency.lockutils [req-184a9df8-439c-4d79-a830-d41fe039dad1 req-70d554eb-689c-4844-b672-6cb1a70737a2 service nova] Acquiring lock "refresh_cache-7f341d80-450a-47b8-b26d-15f2ce3e378c" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.078195] env[63088]: DEBUG oslo_concurrency.lockutils [req-184a9df8-439c-4d79-a830-d41fe039dad1 req-70d554eb-689c-4844-b672-6cb1a70737a2 service nova] Acquired lock "refresh_cache-7f341d80-450a-47b8-b26d-15f2ce3e378c" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.078390] env[63088]: DEBUG nova.network.neutron [req-184a9df8-439c-4d79-a830-d41fe039dad1 req-70d554eb-689c-4844-b672-6cb1a70737a2 service nova] [instance: 7f341d80-450a-47b8-b26d-15f2ce3e378c] Refreshing network info cache for port 711f2731-680e-4328-97c8-0a2d1bbf79eb {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 844.080093] env[63088]: DEBUG oslo_concurrency.lockutils [req-1fffb8c2-a9b0-4f4b-affe-174c62ff656f req-2bbba38c-19b2-4181-9690-aa8db145e658 service nova] Acquired lock "refresh_cache-07fd3e6f-e0af-467c-9039-238be5c58d25" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.080322] env[63088]: DEBUG nova.network.neutron [req-1fffb8c2-a9b0-4f4b-affe-174c62ff656f req-2bbba38c-19b2-4181-9690-aa8db145e658 service nova] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Refreshing network info cache for port 0c182796-7fb6-40ab-ba27-a43f0700dee1 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 844.091678] env[63088]: DEBUG oslo_vmware.api [None req-b548d275-1b16-4d7d-aec5-f1421ffffdea tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Task: {'id': task-1284972, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.251771} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.091678] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-b548d275-1b16-4d7d-aec5-f1421ffffdea tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 844.091678] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-b548d275-1b16-4d7d-aec5-f1421ffffdea tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] [instance: 2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68] Deleted contents of the VM from datastore datastore2 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 844.093308] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-b548d275-1b16-4d7d-aec5-f1421ffffdea tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] [instance: 2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 844.093308] env[63088]: INFO nova.compute.manager [None req-b548d275-1b16-4d7d-aec5-f1421ffffdea tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] [instance: 2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68] Took 1.14 seconds to destroy the instance on the hypervisor. [ 844.093308] env[63088]: DEBUG oslo.service.loopingcall [None req-b548d275-1b16-4d7d-aec5-f1421ffffdea tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 844.093694] env[63088]: DEBUG nova.compute.manager [-] [instance: 2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 844.093694] env[63088]: DEBUG nova.network.neutron [-] [instance: 2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 844.139691] env[63088]: DEBUG oslo_vmware.api [None req-8102b1ea-81e5-4196-b4f7-a4fd361924a7 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Task: {'id': task-1284969, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.249647] env[63088]: DEBUG oslo_vmware.api [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52cc0bbc-8cbc-04ce-b83a-7e9bfcf14b9a, 'name': SearchDatastore_Task, 'duration_secs': 0.016416} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.250340] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4cc00ffd-8aee-4754-b14b-724adcc1ea90 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.256352] env[63088]: DEBUG oslo_vmware.api [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Waiting for the task: (returnval){ [ 844.256352] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]529abd9f-5abd-85ab-1651-4cdb72d7dd05" [ 844.256352] env[63088]: _type = "Task" [ 844.256352] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.264892] env[63088]: DEBUG oslo_vmware.api [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]529abd9f-5abd-85ab-1651-4cdb72d7dd05, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.396644] env[63088]: DEBUG nova.network.neutron [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 8b564e29-96cf-4abf-963d-142b413fb464] Updating instance_info_cache with network_info: [{"id": "7a0cbd40-cac8-4bb6-944c-486989457770", "address": "fa:16:3e:c2:dd:41", "network": {"id": "2bbaf351-fe63-46ba-ba7e-0f432c0a1bfb", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1438884226-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7bff1382b9694df08133c88a5fe783a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cca1f087-01e1-49ca-831b-5c51478a5d60", "external-id": "nsx-vlan-transportzone-439", "segmentation_id": 439, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7a0cbd40-ca", "ovs_interfaceid": "7a0cbd40-cac8-4bb6-944c-486989457770", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 844.412192] env[63088]: DEBUG nova.compute.manager [req-d75e1035-cc37-4826-8728-04a8fbbb18df req-9cb1bcc3-0f93-4a9a-9cfe-c2652e20f4b9 service nova] [instance: 8b564e29-96cf-4abf-963d-142b413fb464] Received event network-vif-plugged-7a0cbd40-cac8-4bb6-944c-486989457770 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 844.412421] env[63088]: DEBUG oslo_concurrency.lockutils [req-d75e1035-cc37-4826-8728-04a8fbbb18df req-9cb1bcc3-0f93-4a9a-9cfe-c2652e20f4b9 service nova] Acquiring lock "8b564e29-96cf-4abf-963d-142b413fb464-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.412622] env[63088]: DEBUG oslo_concurrency.lockutils [req-d75e1035-cc37-4826-8728-04a8fbbb18df req-9cb1bcc3-0f93-4a9a-9cfe-c2652e20f4b9 service nova] Lock "8b564e29-96cf-4abf-963d-142b413fb464-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.412788] env[63088]: DEBUG oslo_concurrency.lockutils [req-d75e1035-cc37-4826-8728-04a8fbbb18df req-9cb1bcc3-0f93-4a9a-9cfe-c2652e20f4b9 service nova] Lock "8b564e29-96cf-4abf-963d-142b413fb464-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.412953] env[63088]: DEBUG nova.compute.manager [req-d75e1035-cc37-4826-8728-04a8fbbb18df req-9cb1bcc3-0f93-4a9a-9cfe-c2652e20f4b9 service nova] [instance: 8b564e29-96cf-4abf-963d-142b413fb464] No waiting events found dispatching network-vif-plugged-7a0cbd40-cac8-4bb6-944c-486989457770 {{(pid=63088) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 844.413187] env[63088]: WARNING nova.compute.manager [req-d75e1035-cc37-4826-8728-04a8fbbb18df req-9cb1bcc3-0f93-4a9a-9cfe-c2652e20f4b9 service nova] [instance: 8b564e29-96cf-4abf-963d-142b413fb464] Received unexpected event network-vif-plugged-7a0cbd40-cac8-4bb6-944c-486989457770 for instance with vm_state building and task_state spawning. [ 844.413361] env[63088]: DEBUG nova.compute.manager [req-d75e1035-cc37-4826-8728-04a8fbbb18df req-9cb1bcc3-0f93-4a9a-9cfe-c2652e20f4b9 service nova] [instance: 8b564e29-96cf-4abf-963d-142b413fb464] Received event network-changed-7a0cbd40-cac8-4bb6-944c-486989457770 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 844.413519] env[63088]: DEBUG nova.compute.manager [req-d75e1035-cc37-4826-8728-04a8fbbb18df req-9cb1bcc3-0f93-4a9a-9cfe-c2652e20f4b9 service nova] [instance: 8b564e29-96cf-4abf-963d-142b413fb464] Refreshing instance network info cache due to event network-changed-7a0cbd40-cac8-4bb6-944c-486989457770. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 844.413697] env[63088]: DEBUG oslo_concurrency.lockutils [req-d75e1035-cc37-4826-8728-04a8fbbb18df req-9cb1bcc3-0f93-4a9a-9cfe-c2652e20f4b9 service nova] Acquiring lock "refresh_cache-8b564e29-96cf-4abf-963d-142b413fb464" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.433320] env[63088]: DEBUG nova.scheduler.client.report [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 844.439095] env[63088]: DEBUG nova.compute.manager [req-ea198cda-9a18-474d-89fe-1e89644ea8af req-6cbc791e-638d-4d83-8361-436c83d2a831 service nova] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Received event network-changed-0c182796-7fb6-40ab-ba27-a43f0700dee1 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 844.439306] env[63088]: DEBUG nova.compute.manager [req-ea198cda-9a18-474d-89fe-1e89644ea8af req-6cbc791e-638d-4d83-8361-436c83d2a831 service nova] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Refreshing instance network info cache due to event network-changed-0c182796-7fb6-40ab-ba27-a43f0700dee1. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 844.439472] env[63088]: DEBUG oslo_concurrency.lockutils [req-ea198cda-9a18-474d-89fe-1e89644ea8af req-6cbc791e-638d-4d83-8361-436c83d2a831 service nova] Acquiring lock "refresh_cache-07fd3e6f-e0af-467c-9039-238be5c58d25" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.475022] env[63088]: DEBUG oslo_vmware.api [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': task-1284973, 'name': Rename_Task, 'duration_secs': 0.139487} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.475022] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 00925f94-dbf9-453c-a124-d8434679aedf] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 844.475022] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8e74cee6-fd02-41ab-b128-5290e2bc53ba {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.479608] env[63088]: DEBUG oslo_vmware.api [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Waiting for the task: (returnval){ [ 844.479608] env[63088]: value = "task-1284974" [ 844.479608] env[63088]: _type = "Task" [ 844.479608] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.490884] env[63088]: DEBUG oslo_vmware.api [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': task-1284974, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.581610] env[63088]: DEBUG nova.compute.manager [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 844.615249] env[63088]: DEBUG nova.virt.hardware [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 844.615490] env[63088]: DEBUG nova.virt.hardware [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 844.615647] env[63088]: DEBUG nova.virt.hardware [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 844.615826] env[63088]: DEBUG nova.virt.hardware [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 844.615971] env[63088]: DEBUG nova.virt.hardware [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 844.616132] env[63088]: DEBUG nova.virt.hardware [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 844.616336] env[63088]: DEBUG nova.virt.hardware [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 844.616634] env[63088]: DEBUG nova.virt.hardware [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 844.616702] env[63088]: DEBUG nova.virt.hardware [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 844.616814] env[63088]: DEBUG nova.virt.hardware [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 844.616990] env[63088]: DEBUG nova.virt.hardware [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 844.617861] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61e65942-63d4-4b84-a428-585034c31e12 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.625736] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03577b28-98a0-46aa-b49b-016f6c7ac5cc {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.649440] env[63088]: DEBUG oslo_vmware.api [None req-8102b1ea-81e5-4196-b4f7-a4fd361924a7 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Task: {'id': task-1284969, 'name': PowerOffVM_Task, 'duration_secs': 1.160335} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.649728] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-8102b1ea-81e5-4196-b4f7-a4fd361924a7 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] [instance: b159649b-ebd5-4c7d-9074-dc0e6395a947] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 844.649875] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-8102b1ea-81e5-4196-b4f7-a4fd361924a7 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] [instance: b159649b-ebd5-4c7d-9074-dc0e6395a947] Unregistering the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 844.650142] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-20083ae0-96b6-453c-b685-d0db63aaaaa1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.677930] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-8102b1ea-81e5-4196-b4f7-a4fd361924a7 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] [instance: b159649b-ebd5-4c7d-9074-dc0e6395a947] Unregistered the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 844.677930] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-8102b1ea-81e5-4196-b4f7-a4fd361924a7 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] [instance: b159649b-ebd5-4c7d-9074-dc0e6395a947] Deleting contents of the VM from datastore datastore2 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 844.677930] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-8102b1ea-81e5-4196-b4f7-a4fd361924a7 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Deleting the datastore file [datastore2] b159649b-ebd5-4c7d-9074-dc0e6395a947 {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 844.678221] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-be6d8fd2-ca2e-4a67-8bb9-98defe65a54d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.686262] env[63088]: DEBUG oslo_vmware.api [None req-8102b1ea-81e5-4196-b4f7-a4fd361924a7 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Waiting for the task: (returnval){ [ 844.686262] env[63088]: value = "task-1284976" [ 844.686262] env[63088]: _type = "Task" [ 844.686262] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.696854] env[63088]: DEBUG oslo_vmware.api [None req-8102b1ea-81e5-4196-b4f7-a4fd361924a7 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Task: {'id': task-1284976, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.766419] env[63088]: DEBUG oslo_vmware.api [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]529abd9f-5abd-85ab-1651-4cdb72d7dd05, 'name': SearchDatastore_Task, 'duration_secs': 0.021716} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.768826] env[63088]: DEBUG oslo_concurrency.lockutils [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.769143] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] 7f341d80-450a-47b8-b26d-15f2ce3e378c/7f341d80-450a-47b8-b26d-15f2ce3e378c.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 844.769438] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-31b3be00-fd75-4998-9f10-21e044f559d0 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.775944] env[63088]: DEBUG oslo_vmware.api [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Waiting for the task: (returnval){ [ 844.775944] env[63088]: value = "task-1284977" [ 844.775944] env[63088]: _type = "Task" [ 844.775944] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.786283] env[63088]: DEBUG oslo_vmware.api [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': task-1284977, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.846984] env[63088]: DEBUG nova.network.neutron [req-184a9df8-439c-4d79-a830-d41fe039dad1 req-70d554eb-689c-4844-b672-6cb1a70737a2 service nova] [instance: 7f341d80-450a-47b8-b26d-15f2ce3e378c] Updated VIF entry in instance network info cache for port 711f2731-680e-4328-97c8-0a2d1bbf79eb. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 844.847401] env[63088]: DEBUG nova.network.neutron [req-184a9df8-439c-4d79-a830-d41fe039dad1 req-70d554eb-689c-4844-b672-6cb1a70737a2 service nova] [instance: 7f341d80-450a-47b8-b26d-15f2ce3e378c] Updating instance_info_cache with network_info: [{"id": "711f2731-680e-4328-97c8-0a2d1bbf79eb", "address": "fa:16:3e:46:39:13", "network": {"id": "79155637-c844-4012-a77a-d04f9ec9ab18", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-2144942540-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e27be174c0b3400193f9ae64a242da1a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap711f2731-68", "ovs_interfaceid": "711f2731-680e-4328-97c8-0a2d1bbf79eb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 844.888183] env[63088]: DEBUG nova.network.neutron [req-1fffb8c2-a9b0-4f4b-affe-174c62ff656f req-2bbba38c-19b2-4181-9690-aa8db145e658 service nova] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Updated VIF entry in instance network info cache for port 0c182796-7fb6-40ab-ba27-a43f0700dee1. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 844.888583] env[63088]: DEBUG nova.network.neutron [req-1fffb8c2-a9b0-4f4b-affe-174c62ff656f req-2bbba38c-19b2-4181-9690-aa8db145e658 service nova] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Updating instance_info_cache with network_info: [{"id": "0c182796-7fb6-40ab-ba27-a43f0700dee1", "address": "fa:16:3e:63:25:2a", "network": {"id": "de116097-3737-4dcf-972c-88c49fa8995d", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-580707267-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "937eba1a40c84534b279d5798e234625", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3a80436-f7a9-431a-acec-aca3d76e3f9b", "external-id": "cl2-zone-339", "segmentation_id": 339, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c182796-7f", "ovs_interfaceid": "0c182796-7fb6-40ab-ba27-a43f0700dee1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 844.901408] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Releasing lock "refresh_cache-8b564e29-96cf-4abf-963d-142b413fb464" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.901739] env[63088]: DEBUG nova.compute.manager [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 8b564e29-96cf-4abf-963d-142b413fb464] Instance network_info: |[{"id": "7a0cbd40-cac8-4bb6-944c-486989457770", "address": "fa:16:3e:c2:dd:41", "network": {"id": "2bbaf351-fe63-46ba-ba7e-0f432c0a1bfb", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1438884226-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7bff1382b9694df08133c88a5fe783a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cca1f087-01e1-49ca-831b-5c51478a5d60", "external-id": "nsx-vlan-transportzone-439", "segmentation_id": 439, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7a0cbd40-ca", "ovs_interfaceid": "7a0cbd40-cac8-4bb6-944c-486989457770", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 844.902081] env[63088]: DEBUG oslo_concurrency.lockutils [req-d75e1035-cc37-4826-8728-04a8fbbb18df req-9cb1bcc3-0f93-4a9a-9cfe-c2652e20f4b9 service nova] Acquired lock "refresh_cache-8b564e29-96cf-4abf-963d-142b413fb464" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.902279] env[63088]: DEBUG nova.network.neutron [req-d75e1035-cc37-4826-8728-04a8fbbb18df req-9cb1bcc3-0f93-4a9a-9cfe-c2652e20f4b9 service nova] [instance: 8b564e29-96cf-4abf-963d-142b413fb464] Refreshing network info cache for port 7a0cbd40-cac8-4bb6-944c-486989457770 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 844.906347] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 8b564e29-96cf-4abf-963d-142b413fb464] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c2:dd:41', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cca1f087-01e1-49ca-831b-5c51478a5d60', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7a0cbd40-cac8-4bb6-944c-486989457770', 'vif_model': 'vmxnet3'}] {{(pid=63088) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 844.912040] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Creating folder: Project (7bff1382b9694df08133c88a5fe783a5). Parent ref: group-v275816. {{(pid=63088) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 844.912392] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4ab6fe2a-dfb1-4e55-a67e-105a6cfc95d8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.925021] env[63088]: DEBUG nova.network.neutron [-] [instance: 2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 844.925021] env[63088]: INFO nova.virt.vmwareapi.vm_util [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Created folder: Project (7bff1382b9694df08133c88a5fe783a5) in parent group-v275816. [ 844.925021] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Creating folder: Instances. Parent ref: group-v275872. {{(pid=63088) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 844.925021] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2380e3ba-95af-4d13-a4f3-482186757b0e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.933322] env[63088]: INFO nova.virt.vmwareapi.vm_util [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Created folder: Instances in parent group-v275872. [ 844.933635] env[63088]: DEBUG oslo.service.loopingcall [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 844.933878] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8b564e29-96cf-4abf-963d-142b413fb464] Creating VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 844.934137] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8ddbf7d0-c253-46bf-b935-aa6fc8d5d0d9 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.950185] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.403s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.950660] env[63088]: DEBUG nova.compute.manager [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 844.953368] env[63088]: DEBUG oslo_concurrency.lockutils [None req-049e65c1-0c59-41b7-8448-5a3e6eb7d38f tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.538s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.953601] env[63088]: DEBUG nova.objects.instance [None req-049e65c1-0c59-41b7-8448-5a3e6eb7d38f tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lazy-loading 'resources' on Instance uuid e845f60b-004f-4a5c-813e-0c183e99eba6 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 844.962019] env[63088]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 844.962019] env[63088]: value = "task-1284980" [ 844.962019] env[63088]: _type = "Task" [ 844.962019] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.971617] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1284980, 'name': CreateVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.990700] env[63088]: DEBUG oslo_vmware.api [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': task-1284974, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.196961] env[63088]: DEBUG oslo_vmware.api [None req-8102b1ea-81e5-4196-b4f7-a4fd361924a7 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Task: {'id': task-1284976, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.182346} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.198217] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-8102b1ea-81e5-4196-b4f7-a4fd361924a7 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 845.198217] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-8102b1ea-81e5-4196-b4f7-a4fd361924a7 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] [instance: b159649b-ebd5-4c7d-9074-dc0e6395a947] Deleted contents of the VM from datastore datastore2 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 845.198217] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-8102b1ea-81e5-4196-b4f7-a4fd361924a7 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] [instance: b159649b-ebd5-4c7d-9074-dc0e6395a947] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 845.198217] env[63088]: INFO nova.compute.manager [None req-8102b1ea-81e5-4196-b4f7-a4fd361924a7 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] [instance: b159649b-ebd5-4c7d-9074-dc0e6395a947] Took 2.09 seconds to destroy the instance on the hypervisor. [ 845.198461] env[63088]: DEBUG oslo.service.loopingcall [None req-8102b1ea-81e5-4196-b4f7-a4fd361924a7 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 845.198663] env[63088]: DEBUG nova.compute.manager [-] [instance: b159649b-ebd5-4c7d-9074-dc0e6395a947] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 845.198778] env[63088]: DEBUG nova.network.neutron [-] [instance: b159649b-ebd5-4c7d-9074-dc0e6395a947] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 845.220332] env[63088]: DEBUG nova.network.neutron [-] [instance: b159649b-ebd5-4c7d-9074-dc0e6395a947] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 845.289836] env[63088]: DEBUG oslo_vmware.api [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': task-1284977, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.307584] env[63088]: DEBUG oslo_concurrency.lockutils [None req-ad047aea-ed6a-4a24-b3ca-46aa3078b110 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Acquiring lock "07fd3e6f-e0af-467c-9039-238be5c58d25" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 845.307884] env[63088]: DEBUG oslo_concurrency.lockutils [None req-ad047aea-ed6a-4a24-b3ca-46aa3078b110 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Lock "07fd3e6f-e0af-467c-9039-238be5c58d25" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 845.308098] env[63088]: DEBUG oslo_concurrency.lockutils [None req-ad047aea-ed6a-4a24-b3ca-46aa3078b110 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Acquiring lock "07fd3e6f-e0af-467c-9039-238be5c58d25-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 845.308287] env[63088]: DEBUG oslo_concurrency.lockutils [None req-ad047aea-ed6a-4a24-b3ca-46aa3078b110 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Lock "07fd3e6f-e0af-467c-9039-238be5c58d25-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 845.308483] env[63088]: DEBUG oslo_concurrency.lockutils [None req-ad047aea-ed6a-4a24-b3ca-46aa3078b110 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Lock "07fd3e6f-e0af-467c-9039-238be5c58d25-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 845.311055] env[63088]: INFO nova.compute.manager [None req-ad047aea-ed6a-4a24-b3ca-46aa3078b110 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Terminating instance [ 845.313155] env[63088]: DEBUG nova.compute.manager [None req-ad047aea-ed6a-4a24-b3ca-46aa3078b110 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 845.313318] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-ad047aea-ed6a-4a24-b3ca-46aa3078b110 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 845.314259] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f44de09-6bf6-4271-8658-8a2b4bd49be3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.322018] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad047aea-ed6a-4a24-b3ca-46aa3078b110 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 845.322173] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-473265ae-18de-469c-b597-44b29f2b0191 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.328281] env[63088]: DEBUG oslo_vmware.api [None req-ad047aea-ed6a-4a24-b3ca-46aa3078b110 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Waiting for the task: (returnval){ [ 845.328281] env[63088]: value = "task-1284981" [ 845.328281] env[63088]: _type = "Task" [ 845.328281] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.336607] env[63088]: DEBUG oslo_vmware.api [None req-ad047aea-ed6a-4a24-b3ca-46aa3078b110 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Task: {'id': task-1284981, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.350625] env[63088]: DEBUG oslo_concurrency.lockutils [req-184a9df8-439c-4d79-a830-d41fe039dad1 req-70d554eb-689c-4844-b672-6cb1a70737a2 service nova] Releasing lock "refresh_cache-7f341d80-450a-47b8-b26d-15f2ce3e378c" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.391752] env[63088]: DEBUG oslo_concurrency.lockutils [req-1fffb8c2-a9b0-4f4b-affe-174c62ff656f req-2bbba38c-19b2-4181-9690-aa8db145e658 service nova] Releasing lock "refresh_cache-07fd3e6f-e0af-467c-9039-238be5c58d25" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.392266] env[63088]: DEBUG oslo_concurrency.lockutils [req-ea198cda-9a18-474d-89fe-1e89644ea8af req-6cbc791e-638d-4d83-8361-436c83d2a831 service nova] Acquired lock "refresh_cache-07fd3e6f-e0af-467c-9039-238be5c58d25" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.392471] env[63088]: DEBUG nova.network.neutron [req-ea198cda-9a18-474d-89fe-1e89644ea8af req-6cbc791e-638d-4d83-8361-436c83d2a831 service nova] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Refreshing network info cache for port 0c182796-7fb6-40ab-ba27-a43f0700dee1 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 845.426077] env[63088]: INFO nova.compute.manager [-] [instance: 2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68] Took 1.33 seconds to deallocate network for instance. [ 845.457306] env[63088]: DEBUG nova.compute.utils [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 845.465471] env[63088]: DEBUG nova.compute.manager [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 845.465471] env[63088]: DEBUG nova.network.neutron [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 845.477950] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1284980, 'name': CreateVM_Task, 'duration_secs': 0.395284} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.478133] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8b564e29-96cf-4abf-963d-142b413fb464] Created VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 845.478832] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 845.479620] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.480079] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 845.482941] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e1773ad8-2a2d-48bf-bb6a-c3c9fbf3fab6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.491290] env[63088]: DEBUG oslo_vmware.api [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Waiting for the task: (returnval){ [ 845.491290] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52a389ec-5a6c-4b0e-8eb9-51b891186969" [ 845.491290] env[63088]: _type = "Task" [ 845.491290] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.494763] env[63088]: DEBUG oslo_vmware.api [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': task-1284974, 'name': PowerOnVM_Task, 'duration_secs': 0.554547} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.498090] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 00925f94-dbf9-453c-a124-d8434679aedf] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 845.498305] env[63088]: INFO nova.compute.manager [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 00925f94-dbf9-453c-a124-d8434679aedf] Took 8.45 seconds to spawn the instance on the hypervisor. [ 845.498487] env[63088]: DEBUG nova.compute.manager [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 00925f94-dbf9-453c-a124-d8434679aedf] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 845.501619] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79470a27-b908-476c-b4a4-3ed1d6057df2 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.511284] env[63088]: DEBUG oslo_vmware.api [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52a389ec-5a6c-4b0e-8eb9-51b891186969, 'name': SearchDatastore_Task, 'duration_secs': 0.008572} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.513968] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.514239] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 8b564e29-96cf-4abf-963d-142b413fb464] Processing image 1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 845.514526] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 845.514800] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.514969] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 845.523853] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fc7f6829-7036-45f8-8b97-59c8f0402f69 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.535034] env[63088]: DEBUG nova.policy [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0d4202df2fce486a96fea968d0189398', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7ed1e04b27ee464eb9a72866d8327352', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 845.541792] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 845.541987] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 845.542755] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-41c1673b-c630-4495-95b9-92331cfe4e94 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.550924] env[63088]: DEBUG oslo_vmware.api [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Waiting for the task: (returnval){ [ 845.550924] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52f1dece-2096-446b-deed-ba1d8b18e49e" [ 845.550924] env[63088]: _type = "Task" [ 845.550924] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.565045] env[63088]: DEBUG oslo_vmware.api [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52f1dece-2096-446b-deed-ba1d8b18e49e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.698019] env[63088]: DEBUG nova.network.neutron [req-d75e1035-cc37-4826-8728-04a8fbbb18df req-9cb1bcc3-0f93-4a9a-9cfe-c2652e20f4b9 service nova] [instance: 8b564e29-96cf-4abf-963d-142b413fb464] Updated VIF entry in instance network info cache for port 7a0cbd40-cac8-4bb6-944c-486989457770. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 845.698370] env[63088]: DEBUG nova.network.neutron [req-d75e1035-cc37-4826-8728-04a8fbbb18df req-9cb1bcc3-0f93-4a9a-9cfe-c2652e20f4b9 service nova] [instance: 8b564e29-96cf-4abf-963d-142b413fb464] Updating instance_info_cache with network_info: [{"id": "7a0cbd40-cac8-4bb6-944c-486989457770", "address": "fa:16:3e:c2:dd:41", "network": {"id": "2bbaf351-fe63-46ba-ba7e-0f432c0a1bfb", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1438884226-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7bff1382b9694df08133c88a5fe783a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cca1f087-01e1-49ca-831b-5c51478a5d60", "external-id": "nsx-vlan-transportzone-439", "segmentation_id": 439, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7a0cbd40-ca", "ovs_interfaceid": "7a0cbd40-cac8-4bb6-944c-486989457770", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 845.723346] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c39e0557-36eb-432c-8e51-8136da99476a tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Acquiring lock "1d421bf0-925d-408c-a612-b659da92799f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 845.723593] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c39e0557-36eb-432c-8e51-8136da99476a tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Lock "1d421bf0-925d-408c-a612-b659da92799f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 845.723807] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c39e0557-36eb-432c-8e51-8136da99476a tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Acquiring lock "1d421bf0-925d-408c-a612-b659da92799f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 845.725508] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c39e0557-36eb-432c-8e51-8136da99476a tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Lock "1d421bf0-925d-408c-a612-b659da92799f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 845.725508] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c39e0557-36eb-432c-8e51-8136da99476a tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Lock "1d421bf0-925d-408c-a612-b659da92799f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 845.726886] env[63088]: DEBUG nova.network.neutron [-] [instance: b159649b-ebd5-4c7d-9074-dc0e6395a947] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 845.728326] env[63088]: INFO nova.compute.manager [None req-c39e0557-36eb-432c-8e51-8136da99476a tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] [instance: 1d421bf0-925d-408c-a612-b659da92799f] Terminating instance [ 845.731903] env[63088]: DEBUG nova.compute.manager [None req-c39e0557-36eb-432c-8e51-8136da99476a tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] [instance: 1d421bf0-925d-408c-a612-b659da92799f] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 845.734610] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-c39e0557-36eb-432c-8e51-8136da99476a tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] [instance: 1d421bf0-925d-408c-a612-b659da92799f] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 845.734610] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-433b53a9-8859-4df8-bf4a-fcad41eeb8e9 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.744375] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-c39e0557-36eb-432c-8e51-8136da99476a tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] [instance: 1d421bf0-925d-408c-a612-b659da92799f] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 845.744633] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c10683a4-c6a8-491d-9c5e-43347e96f027 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.751340] env[63088]: DEBUG oslo_vmware.api [None req-c39e0557-36eb-432c-8e51-8136da99476a tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Waiting for the task: (returnval){ [ 845.751340] env[63088]: value = "task-1284982" [ 845.751340] env[63088]: _type = "Task" [ 845.751340] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.762338] env[63088]: DEBUG oslo_vmware.api [None req-c39e0557-36eb-432c-8e51-8136da99476a tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Task: {'id': task-1284982, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.783073] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf1c20fa-512e-4bb3-a79a-2a480386976e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.791143] env[63088]: DEBUG oslo_vmware.api [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': task-1284977, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.699516} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.792358] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] 7f341d80-450a-47b8-b26d-15f2ce3e378c/7f341d80-450a-47b8-b26d-15f2ce3e378c.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 845.792573] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 7f341d80-450a-47b8-b26d-15f2ce3e378c] Extending root virtual disk to 1048576 {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 845.792864] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d9f2c845-577b-4ec3-a1ba-4070ab71fb6f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.795336] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-646d3366-71fd-42c0-b393-40df4dd7745c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.831047] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fda9b2a-d53f-470f-925a-954ed24486ea {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.834174] env[63088]: DEBUG oslo_vmware.api [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Waiting for the task: (returnval){ [ 845.834174] env[63088]: value = "task-1284983" [ 845.834174] env[63088]: _type = "Task" [ 845.834174] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.845347] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8464f712-18e5-4210-acd7-2cb624e82ccd {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.855547] env[63088]: DEBUG oslo_vmware.api [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': task-1284983, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.855800] env[63088]: DEBUG oslo_vmware.api [None req-ad047aea-ed6a-4a24-b3ca-46aa3078b110 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Task: {'id': task-1284981, 'name': PowerOffVM_Task, 'duration_secs': 0.258332} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.856785] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad047aea-ed6a-4a24-b3ca-46aa3078b110 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 845.856785] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-ad047aea-ed6a-4a24-b3ca-46aa3078b110 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Unregistering the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 845.856937] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b70849c7-04a9-4ac4-840e-167b0e5c5473 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.869398] env[63088]: DEBUG nova.compute.provider_tree [None req-049e65c1-0c59-41b7-8448-5a3e6eb7d38f tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 845.929216] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-ad047aea-ed6a-4a24-b3ca-46aa3078b110 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Unregistered the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 845.929461] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-ad047aea-ed6a-4a24-b3ca-46aa3078b110 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Deleting contents of the VM from datastore datastore2 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 845.929644] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-ad047aea-ed6a-4a24-b3ca-46aa3078b110 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Deleting the datastore file [datastore2] 07fd3e6f-e0af-467c-9039-238be5c58d25 {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 845.929989] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-989860eb-9442-46d8-a820-eb7ddf8a0a96 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.934945] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b548d275-1b16-4d7d-aec5-f1421ffffdea tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 845.939369] env[63088]: DEBUG oslo_vmware.api [None req-ad047aea-ed6a-4a24-b3ca-46aa3078b110 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Waiting for the task: (returnval){ [ 845.939369] env[63088]: value = "task-1284985" [ 845.939369] env[63088]: _type = "Task" [ 845.939369] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.950025] env[63088]: DEBUG oslo_vmware.api [None req-ad047aea-ed6a-4a24-b3ca-46aa3078b110 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Task: {'id': task-1284985, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.965099] env[63088]: DEBUG nova.compute.manager [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 846.034020] env[63088]: INFO nova.compute.manager [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 00925f94-dbf9-453c-a124-d8434679aedf] Took 23.79 seconds to build instance. [ 846.064270] env[63088]: DEBUG oslo_vmware.api [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52f1dece-2096-446b-deed-ba1d8b18e49e, 'name': SearchDatastore_Task, 'duration_secs': 0.015046} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.065180] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9bbea6b4-8f62-4abd-a183-95ced66025ed {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.070689] env[63088]: DEBUG oslo_vmware.api [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Waiting for the task: (returnval){ [ 846.070689] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52768841-748d-9600-33ee-676f9ad3d097" [ 846.070689] env[63088]: _type = "Task" [ 846.070689] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.078801] env[63088]: DEBUG oslo_vmware.api [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52768841-748d-9600-33ee-676f9ad3d097, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.113472] env[63088]: DEBUG nova.network.neutron [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Successfully updated port: fa4a91d1-df9a-4789-bc5c-a8b95457cd93 {{(pid=63088) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 846.131544] env[63088]: DEBUG nova.network.neutron [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Successfully created port: 6ce9925d-cf59-4de0-aece-da90fb4b9f36 {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 846.171942] env[63088]: DEBUG nova.network.neutron [req-ea198cda-9a18-474d-89fe-1e89644ea8af req-6cbc791e-638d-4d83-8361-436c83d2a831 service nova] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Updated VIF entry in instance network info cache for port 0c182796-7fb6-40ab-ba27-a43f0700dee1. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 846.172322] env[63088]: DEBUG nova.network.neutron [req-ea198cda-9a18-474d-89fe-1e89644ea8af req-6cbc791e-638d-4d83-8361-436c83d2a831 service nova] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Updating instance_info_cache with network_info: [{"id": "0c182796-7fb6-40ab-ba27-a43f0700dee1", "address": "fa:16:3e:63:25:2a", "network": {"id": "de116097-3737-4dcf-972c-88c49fa8995d", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-580707267-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "937eba1a40c84534b279d5798e234625", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3a80436-f7a9-431a-acec-aca3d76e3f9b", "external-id": "cl2-zone-339", "segmentation_id": 339, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c182796-7f", "ovs_interfaceid": "0c182796-7fb6-40ab-ba27-a43f0700dee1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 846.200826] env[63088]: DEBUG oslo_concurrency.lockutils [req-d75e1035-cc37-4826-8728-04a8fbbb18df req-9cb1bcc3-0f93-4a9a-9cfe-c2652e20f4b9 service nova] Releasing lock "refresh_cache-8b564e29-96cf-4abf-963d-142b413fb464" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 846.231083] env[63088]: INFO nova.compute.manager [-] [instance: b159649b-ebd5-4c7d-9074-dc0e6395a947] Took 1.03 seconds to deallocate network for instance. [ 846.262317] env[63088]: DEBUG oslo_vmware.api [None req-c39e0557-36eb-432c-8e51-8136da99476a tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Task: {'id': task-1284982, 'name': PowerOffVM_Task, 'duration_secs': 0.285929} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.263107] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-c39e0557-36eb-432c-8e51-8136da99476a tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] [instance: 1d421bf0-925d-408c-a612-b659da92799f] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 846.263107] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-c39e0557-36eb-432c-8e51-8136da99476a tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] [instance: 1d421bf0-925d-408c-a612-b659da92799f] Unregistering the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 846.263107] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c0c8bf0d-652d-4e2e-bb73-bbf19973e0c8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.347520] env[63088]: DEBUG oslo_vmware.api [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': task-1284983, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.110824} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.347843] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 7f341d80-450a-47b8-b26d-15f2ce3e378c] Extended root virtual disk {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 846.348686] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3830b82-fe28-4229-892c-44b4cc32e79f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.371389] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 7f341d80-450a-47b8-b26d-15f2ce3e378c] Reconfiguring VM instance instance-00000040 to attach disk [datastore1] 7f341d80-450a-47b8-b26d-15f2ce3e378c/7f341d80-450a-47b8-b26d-15f2ce3e378c.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 846.372346] env[63088]: DEBUG nova.scheduler.client.report [None req-049e65c1-0c59-41b7-8448-5a3e6eb7d38f tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 846.375558] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-657eb57d-6fcc-4ec1-afe4-ccfa385646a3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.392175] env[63088]: DEBUG oslo_concurrency.lockutils [None req-049e65c1-0c59-41b7-8448-5a3e6eb7d38f tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.439s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.394344] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.237s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.395886] env[63088]: INFO nova.compute.claims [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 846.405011] env[63088]: DEBUG oslo_vmware.api [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Waiting for the task: (returnval){ [ 846.405011] env[63088]: value = "task-1284987" [ 846.405011] env[63088]: _type = "Task" [ 846.405011] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.413835] env[63088]: DEBUG oslo_vmware.api [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': task-1284987, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.427393] env[63088]: INFO nova.scheduler.client.report [None req-049e65c1-0c59-41b7-8448-5a3e6eb7d38f tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Deleted allocations for instance e845f60b-004f-4a5c-813e-0c183e99eba6 [ 846.453020] env[63088]: DEBUG oslo_vmware.api [None req-ad047aea-ed6a-4a24-b3ca-46aa3078b110 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Task: {'id': task-1284985, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.21595} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.453020] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-ad047aea-ed6a-4a24-b3ca-46aa3078b110 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 846.453020] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-ad047aea-ed6a-4a24-b3ca-46aa3078b110 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Deleted contents of the VM from datastore datastore2 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 846.453020] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-ad047aea-ed6a-4a24-b3ca-46aa3078b110 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 846.453020] env[63088]: INFO nova.compute.manager [None req-ad047aea-ed6a-4a24-b3ca-46aa3078b110 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Took 1.14 seconds to destroy the instance on the hypervisor. [ 846.453549] env[63088]: DEBUG oslo.service.loopingcall [None req-ad047aea-ed6a-4a24-b3ca-46aa3078b110 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 846.453847] env[63088]: DEBUG nova.compute.manager [-] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 846.454051] env[63088]: DEBUG nova.network.neutron [-] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 846.461624] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-c39e0557-36eb-432c-8e51-8136da99476a tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] [instance: 1d421bf0-925d-408c-a612-b659da92799f] Unregistered the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 846.461624] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-c39e0557-36eb-432c-8e51-8136da99476a tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] [instance: 1d421bf0-925d-408c-a612-b659da92799f] Deleting contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 846.461624] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-c39e0557-36eb-432c-8e51-8136da99476a tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Deleting the datastore file [datastore1] 1d421bf0-925d-408c-a612-b659da92799f {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 846.462092] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ecfb2655-078e-4afb-b8e7-7093ada40d36 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.473585] env[63088]: DEBUG oslo_vmware.api [None req-c39e0557-36eb-432c-8e51-8136da99476a tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Waiting for the task: (returnval){ [ 846.473585] env[63088]: value = "task-1284988" [ 846.473585] env[63088]: _type = "Task" [ 846.473585] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.484588] env[63088]: DEBUG oslo_vmware.api [None req-c39e0557-36eb-432c-8e51-8136da99476a tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Task: {'id': task-1284988, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.525009] env[63088]: DEBUG nova.compute.manager [req-81f6fdc0-7696-4663-b7d1-81bf0493b8d7 req-aec04885-00f5-46d6-92d1-dca930f6654a service nova] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Received event network-vif-plugged-fa4a91d1-df9a-4789-bc5c-a8b95457cd93 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 846.525243] env[63088]: DEBUG oslo_concurrency.lockutils [req-81f6fdc0-7696-4663-b7d1-81bf0493b8d7 req-aec04885-00f5-46d6-92d1-dca930f6654a service nova] Acquiring lock "7b6aadb7-e34b-42b7-b69f-370434f5b665-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 846.525446] env[63088]: DEBUG oslo_concurrency.lockutils [req-81f6fdc0-7696-4663-b7d1-81bf0493b8d7 req-aec04885-00f5-46d6-92d1-dca930f6654a service nova] Lock "7b6aadb7-e34b-42b7-b69f-370434f5b665-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.526133] env[63088]: DEBUG oslo_concurrency.lockutils [req-81f6fdc0-7696-4663-b7d1-81bf0493b8d7 req-aec04885-00f5-46d6-92d1-dca930f6654a service nova] Lock "7b6aadb7-e34b-42b7-b69f-370434f5b665-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.526296] env[63088]: DEBUG nova.compute.manager [req-81f6fdc0-7696-4663-b7d1-81bf0493b8d7 req-aec04885-00f5-46d6-92d1-dca930f6654a service nova] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] No waiting events found dispatching network-vif-plugged-fa4a91d1-df9a-4789-bc5c-a8b95457cd93 {{(pid=63088) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 846.526337] env[63088]: WARNING nova.compute.manager [req-81f6fdc0-7696-4663-b7d1-81bf0493b8d7 req-aec04885-00f5-46d6-92d1-dca930f6654a service nova] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Received unexpected event network-vif-plugged-fa4a91d1-df9a-4789-bc5c-a8b95457cd93 for instance with vm_state building and task_state spawning. [ 846.526497] env[63088]: DEBUG nova.compute.manager [req-81f6fdc0-7696-4663-b7d1-81bf0493b8d7 req-aec04885-00f5-46d6-92d1-dca930f6654a service nova] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Received event network-changed-fa4a91d1-df9a-4789-bc5c-a8b95457cd93 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 846.526738] env[63088]: DEBUG nova.compute.manager [req-81f6fdc0-7696-4663-b7d1-81bf0493b8d7 req-aec04885-00f5-46d6-92d1-dca930f6654a service nova] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Refreshing instance network info cache due to event network-changed-fa4a91d1-df9a-4789-bc5c-a8b95457cd93. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 846.526945] env[63088]: DEBUG oslo_concurrency.lockutils [req-81f6fdc0-7696-4663-b7d1-81bf0493b8d7 req-aec04885-00f5-46d6-92d1-dca930f6654a service nova] Acquiring lock "refresh_cache-7b6aadb7-e34b-42b7-b69f-370434f5b665" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 846.527388] env[63088]: DEBUG oslo_concurrency.lockutils [req-81f6fdc0-7696-4663-b7d1-81bf0493b8d7 req-aec04885-00f5-46d6-92d1-dca930f6654a service nova] Acquired lock "refresh_cache-7b6aadb7-e34b-42b7-b69f-370434f5b665" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 846.527388] env[63088]: DEBUG nova.network.neutron [req-81f6fdc0-7696-4663-b7d1-81bf0493b8d7 req-aec04885-00f5-46d6-92d1-dca930f6654a service nova] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Refreshing network info cache for port fa4a91d1-df9a-4789-bc5c-a8b95457cd93 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 846.538920] env[63088]: DEBUG oslo_concurrency.lockutils [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Lock "00925f94-dbf9-453c-a124-d8434679aedf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 55.324s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.546663] env[63088]: DEBUG nova.compute.manager [req-8a4db3ff-471b-4859-8754-4cd0701c927c req-90ac2f17-d80d-454a-b259-a21ec68ff94e service nova] [instance: 2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68] Received event network-vif-deleted-4c03d05a-bb4d-4d2b-bb2c-65f92a25278c {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 846.585015] env[63088]: DEBUG oslo_vmware.api [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52768841-748d-9600-33ee-676f9ad3d097, 'name': SearchDatastore_Task, 'duration_secs': 0.00911} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.585015] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 846.585015] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] 8b564e29-96cf-4abf-963d-142b413fb464/8b564e29-96cf-4abf-963d-142b413fb464.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 846.585015] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-389f3e40-8609-45d0-b368-0bd747962592 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.589777] env[63088]: DEBUG oslo_vmware.api [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Waiting for the task: (returnval){ [ 846.589777] env[63088]: value = "task-1284989" [ 846.589777] env[63088]: _type = "Task" [ 846.589777] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.597810] env[63088]: DEBUG oslo_vmware.api [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1284989, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.616502] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquiring lock "refresh_cache-7b6aadb7-e34b-42b7-b69f-370434f5b665" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 846.675491] env[63088]: DEBUG oslo_concurrency.lockutils [req-ea198cda-9a18-474d-89fe-1e89644ea8af req-6cbc791e-638d-4d83-8361-436c83d2a831 service nova] Releasing lock "refresh_cache-07fd3e6f-e0af-467c-9039-238be5c58d25" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 846.740021] env[63088]: DEBUG oslo_concurrency.lockutils [None req-8102b1ea-81e5-4196-b4f7-a4fd361924a7 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 846.916916] env[63088]: DEBUG oslo_vmware.api [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': task-1284987, 'name': ReconfigVM_Task, 'duration_secs': 0.42397} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.917224] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 7f341d80-450a-47b8-b26d-15f2ce3e378c] Reconfigured VM instance instance-00000040 to attach disk [datastore1] 7f341d80-450a-47b8-b26d-15f2ce3e378c/7f341d80-450a-47b8-b26d-15f2ce3e378c.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 846.917881] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2e3674f8-1de4-450b-a89b-15a25e4fd80e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.926605] env[63088]: DEBUG oslo_vmware.api [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Waiting for the task: (returnval){ [ 846.926605] env[63088]: value = "task-1284990" [ 846.926605] env[63088]: _type = "Task" [ 846.926605] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.940380] env[63088]: DEBUG oslo_vmware.api [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': task-1284990, 'name': Rename_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.941136] env[63088]: DEBUG oslo_concurrency.lockutils [None req-049e65c1-0c59-41b7-8448-5a3e6eb7d38f tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lock "e845f60b-004f-4a5c-813e-0c183e99eba6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 19.035s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.976678] env[63088]: DEBUG nova.compute.manager [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 846.988496] env[63088]: DEBUG oslo_vmware.api [None req-c39e0557-36eb-432c-8e51-8136da99476a tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Task: {'id': task-1284988, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.262175} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.988762] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-c39e0557-36eb-432c-8e51-8136da99476a tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 846.988946] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-c39e0557-36eb-432c-8e51-8136da99476a tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] [instance: 1d421bf0-925d-408c-a612-b659da92799f] Deleted contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 846.989295] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-c39e0557-36eb-432c-8e51-8136da99476a tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] [instance: 1d421bf0-925d-408c-a612-b659da92799f] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 846.989295] env[63088]: INFO nova.compute.manager [None req-c39e0557-36eb-432c-8e51-8136da99476a tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] [instance: 1d421bf0-925d-408c-a612-b659da92799f] Took 1.26 seconds to destroy the instance on the hypervisor. [ 846.989554] env[63088]: DEBUG oslo.service.loopingcall [None req-c39e0557-36eb-432c-8e51-8136da99476a tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 846.990191] env[63088]: DEBUG nova.compute.manager [-] [instance: 1d421bf0-925d-408c-a612-b659da92799f] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 846.990293] env[63088]: DEBUG nova.network.neutron [-] [instance: 1d421bf0-925d-408c-a612-b659da92799f] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 847.003349] env[63088]: DEBUG nova.virt.hardware [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 847.004033] env[63088]: DEBUG nova.virt.hardware [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 847.004033] env[63088]: DEBUG nova.virt.hardware [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 847.004155] env[63088]: DEBUG nova.virt.hardware [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 847.004281] env[63088]: DEBUG nova.virt.hardware [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 847.005788] env[63088]: DEBUG nova.virt.hardware [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 847.005788] env[63088]: DEBUG nova.virt.hardware [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 847.005788] env[63088]: DEBUG nova.virt.hardware [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 847.005788] env[63088]: DEBUG nova.virt.hardware [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 847.005788] env[63088]: DEBUG nova.virt.hardware [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 847.006021] env[63088]: DEBUG nova.virt.hardware [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 847.006249] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-044fd9da-63a7-4a80-8e73-018c6eecf144 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.015024] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02b5aa64-4d15-4c44-b863-71f32edc03d6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.077614] env[63088]: DEBUG nova.network.neutron [req-81f6fdc0-7696-4663-b7d1-81bf0493b8d7 req-aec04885-00f5-46d6-92d1-dca930f6654a service nova] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 847.103917] env[63088]: DEBUG oslo_vmware.api [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1284989, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.218223] env[63088]: DEBUG nova.network.neutron [req-81f6fdc0-7696-4663-b7d1-81bf0493b8d7 req-aec04885-00f5-46d6-92d1-dca930f6654a service nova] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 847.329275] env[63088]: DEBUG nova.network.neutron [-] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 847.436963] env[63088]: DEBUG oslo_vmware.api [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': task-1284990, 'name': Rename_Task, 'duration_secs': 0.467571} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.438675] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 7f341d80-450a-47b8-b26d-15f2ce3e378c] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 847.438675] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-15dfdf27-5090-4e49-9e91-c0b49bd4b606 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.448329] env[63088]: DEBUG oslo_vmware.api [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Waiting for the task: (returnval){ [ 847.448329] env[63088]: value = "task-1284991" [ 847.448329] env[63088]: _type = "Task" [ 847.448329] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.457398] env[63088]: DEBUG oslo_vmware.api [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': task-1284991, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.601942] env[63088]: DEBUG oslo_vmware.api [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1284989, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.685379} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.604984] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] 8b564e29-96cf-4abf-963d-142b413fb464/8b564e29-96cf-4abf-963d-142b413fb464.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 847.605291] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 8b564e29-96cf-4abf-963d-142b413fb464] Extending root virtual disk to 1048576 {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 847.605856] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0fe05ba5-074b-4fb1-b9c3-4eb59135a919 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.613606] env[63088]: DEBUG oslo_vmware.api [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Waiting for the task: (returnval){ [ 847.613606] env[63088]: value = "task-1284992" [ 847.613606] env[63088]: _type = "Task" [ 847.613606] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.627321] env[63088]: DEBUG oslo_vmware.api [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1284992, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.703678] env[63088]: DEBUG nova.compute.manager [req-5fe6b039-e9a4-4e19-8a53-a7be03b80500 req-630a2e8f-5551-43ce-8502-b0718266d4a5 service nova] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Received event network-vif-plugged-6ce9925d-cf59-4de0-aece-da90fb4b9f36 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 847.703950] env[63088]: DEBUG oslo_concurrency.lockutils [req-5fe6b039-e9a4-4e19-8a53-a7be03b80500 req-630a2e8f-5551-43ce-8502-b0718266d4a5 service nova] Acquiring lock "dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 847.704393] env[63088]: DEBUG oslo_concurrency.lockutils [req-5fe6b039-e9a4-4e19-8a53-a7be03b80500 req-630a2e8f-5551-43ce-8502-b0718266d4a5 service nova] Lock "dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 847.704393] env[63088]: DEBUG oslo_concurrency.lockutils [req-5fe6b039-e9a4-4e19-8a53-a7be03b80500 req-630a2e8f-5551-43ce-8502-b0718266d4a5 service nova] Lock "dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 847.704489] env[63088]: DEBUG nova.compute.manager [req-5fe6b039-e9a4-4e19-8a53-a7be03b80500 req-630a2e8f-5551-43ce-8502-b0718266d4a5 service nova] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] No waiting events found dispatching network-vif-plugged-6ce9925d-cf59-4de0-aece-da90fb4b9f36 {{(pid=63088) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 847.704582] env[63088]: WARNING nova.compute.manager [req-5fe6b039-e9a4-4e19-8a53-a7be03b80500 req-630a2e8f-5551-43ce-8502-b0718266d4a5 service nova] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Received unexpected event network-vif-plugged-6ce9925d-cf59-4de0-aece-da90fb4b9f36 for instance with vm_state building and task_state spawning. [ 847.713122] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65eda91b-f98b-496c-9d64-740b5bec3bfc {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.721508] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-600e713c-8b5a-43e7-b6fb-c5785ad096df {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.725473] env[63088]: DEBUG oslo_concurrency.lockutils [req-81f6fdc0-7696-4663-b7d1-81bf0493b8d7 req-aec04885-00f5-46d6-92d1-dca930f6654a service nova] Releasing lock "refresh_cache-7b6aadb7-e34b-42b7-b69f-370434f5b665" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 847.725882] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquired lock "refresh_cache-7b6aadb7-e34b-42b7-b69f-370434f5b665" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 847.726091] env[63088]: DEBUG nova.network.neutron [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 847.757101] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0ecd8ea-68af-4439-9de3-8c08aa8f4da5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.765522] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc281a82-720c-4075-b5d6-92123c829ce0 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.782090] env[63088]: DEBUG nova.compute.provider_tree [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 847.832540] env[63088]: INFO nova.compute.manager [-] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Took 1.38 seconds to deallocate network for instance. [ 847.946202] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquiring lock "0fdb21d1-4111-4ff3-bdc0-e2598298a9a4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 847.946754] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lock "0fdb21d1-4111-4ff3-bdc0-e2598298a9a4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 847.958229] env[63088]: DEBUG oslo_vmware.api [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': task-1284991, 'name': PowerOnVM_Task, 'duration_secs': 0.479} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.958480] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 7f341d80-450a-47b8-b26d-15f2ce3e378c] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 847.958672] env[63088]: INFO nova.compute.manager [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 7f341d80-450a-47b8-b26d-15f2ce3e378c] Took 8.40 seconds to spawn the instance on the hypervisor. [ 847.958845] env[63088]: DEBUG nova.compute.manager [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 7f341d80-450a-47b8-b26d-15f2ce3e378c] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 847.959879] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a7182e6-b058-44b2-9414-fc0261b805a6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.018553] env[63088]: DEBUG nova.network.neutron [-] [instance: 1d421bf0-925d-408c-a612-b659da92799f] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 848.123730] env[63088]: DEBUG oslo_vmware.api [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1284992, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069128} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.123958] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 8b564e29-96cf-4abf-963d-142b413fb464] Extended root virtual disk {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 848.124770] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75cc3398-05fd-40d1-bb31-cb6812c193a5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.148025] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 8b564e29-96cf-4abf-963d-142b413fb464] Reconfiguring VM instance instance-00000041 to attach disk [datastore1] 8b564e29-96cf-4abf-963d-142b413fb464/8b564e29-96cf-4abf-963d-142b413fb464.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 848.148334] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c696db95-d2fd-4ae7-8f19-0313617dce97 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.170075] env[63088]: DEBUG oslo_vmware.api [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Waiting for the task: (returnval){ [ 848.170075] env[63088]: value = "task-1284993" [ 848.170075] env[63088]: _type = "Task" [ 848.170075] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.180893] env[63088]: DEBUG oslo_vmware.api [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1284993, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.187922] env[63088]: DEBUG nova.network.neutron [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Successfully updated port: 6ce9925d-cf59-4de0-aece-da90fb4b9f36 {{(pid=63088) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 848.285453] env[63088]: DEBUG nova.scheduler.client.report [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 848.309390] env[63088]: DEBUG nova.network.neutron [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 848.340069] env[63088]: DEBUG oslo_concurrency.lockutils [None req-ad047aea-ed6a-4a24-b3ca-46aa3078b110 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 848.448839] env[63088]: DEBUG nova.compute.manager [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 0fdb21d1-4111-4ff3-bdc0-e2598298a9a4] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 848.467978] env[63088]: DEBUG nova.network.neutron [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Updating instance_info_cache with network_info: [{"id": "fa4a91d1-df9a-4789-bc5c-a8b95457cd93", "address": "fa:16:3e:7e:c2:c7", "network": {"id": "dfcbe153-bc01-4362-9247-de9f0b1c847b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-841757397-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dace8b5181b84623b08f903d12dfd31e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c68b7663-4f0e-47f0-ac7f-40c6d952f7bb", "external-id": "nsx-vlan-transportzone-696", "segmentation_id": 696, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfa4a91d1-df", "ovs_interfaceid": "fa4a91d1-df9a-4789-bc5c-a8b95457cd93", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 848.479110] env[63088]: INFO nova.compute.manager [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 7f341d80-450a-47b8-b26d-15f2ce3e378c] Took 22.25 seconds to build instance. [ 848.521140] env[63088]: INFO nova.compute.manager [-] [instance: 1d421bf0-925d-408c-a612-b659da92799f] Took 1.53 seconds to deallocate network for instance. [ 848.551166] env[63088]: DEBUG nova.compute.manager [req-a3d6bbb4-2df7-4511-85b6-a75e44638117 req-a88f7a96-f32b-4449-a5de-c888511175e9 service nova] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Received event network-changed-6ce9925d-cf59-4de0-aece-da90fb4b9f36 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 848.552208] env[63088]: DEBUG nova.compute.manager [req-a3d6bbb4-2df7-4511-85b6-a75e44638117 req-a88f7a96-f32b-4449-a5de-c888511175e9 service nova] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Refreshing instance network info cache due to event network-changed-6ce9925d-cf59-4de0-aece-da90fb4b9f36. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 848.552553] env[63088]: DEBUG oslo_concurrency.lockutils [req-a3d6bbb4-2df7-4511-85b6-a75e44638117 req-a88f7a96-f32b-4449-a5de-c888511175e9 service nova] Acquiring lock "refresh_cache-dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 848.552680] env[63088]: DEBUG oslo_concurrency.lockutils [req-a3d6bbb4-2df7-4511-85b6-a75e44638117 req-a88f7a96-f32b-4449-a5de-c888511175e9 service nova] Acquired lock "refresh_cache-dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 848.552833] env[63088]: DEBUG nova.network.neutron [req-a3d6bbb4-2df7-4511-85b6-a75e44638117 req-a88f7a96-f32b-4449-a5de-c888511175e9 service nova] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Refreshing network info cache for port 6ce9925d-cf59-4de0-aece-da90fb4b9f36 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 848.579939] env[63088]: DEBUG nova.compute.manager [req-49619dfa-de5e-4420-b1f8-b69216ddc334 req-4e6754aa-36e4-4b2d-b253-c663a4976c78 service nova] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Received event network-vif-deleted-0c182796-7fb6-40ab-ba27-a43f0700dee1 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 848.580174] env[63088]: DEBUG nova.compute.manager [req-49619dfa-de5e-4420-b1f8-b69216ddc334 req-4e6754aa-36e4-4b2d-b253-c663a4976c78 service nova] [instance: 1d421bf0-925d-408c-a612-b659da92799f] Received event network-vif-deleted-1c83da13-18b2-45c9-bc41-4d45e17606d5 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 848.680593] env[63088]: DEBUG oslo_vmware.api [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1284993, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.690590] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Acquiring lock "refresh_cache-dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 848.791285] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.397s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 848.791825] env[63088]: DEBUG nova.compute.manager [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 848.795045] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a26b18b9-d003-4a4b-9d6c-96d60267d634 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.365s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 848.795883] env[63088]: DEBUG nova.objects.instance [None req-a26b18b9-d003-4a4b-9d6c-96d60267d634 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Lazy-loading 'resources' on Instance uuid 4f6e2d99-7300-42e6-8abf-362ddd4e8c6d {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 848.970889] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Releasing lock "refresh_cache-7b6aadb7-e34b-42b7-b69f-370434f5b665" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 848.971553] env[63088]: DEBUG nova.compute.manager [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Instance network_info: |[{"id": "fa4a91d1-df9a-4789-bc5c-a8b95457cd93", "address": "fa:16:3e:7e:c2:c7", "network": {"id": "dfcbe153-bc01-4362-9247-de9f0b1c847b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-841757397-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dace8b5181b84623b08f903d12dfd31e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c68b7663-4f0e-47f0-ac7f-40c6d952f7bb", "external-id": "nsx-vlan-transportzone-696", "segmentation_id": 696, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfa4a91d1-df", "ovs_interfaceid": "fa4a91d1-df9a-4789-bc5c-a8b95457cd93", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 848.971909] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7e:c2:c7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c68b7663-4f0e-47f0-ac7f-40c6d952f7bb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fa4a91d1-df9a-4789-bc5c-a8b95457cd93', 'vif_model': 'vmxnet3'}] {{(pid=63088) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 848.979500] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Creating folder: Project (dace8b5181b84623b08f903d12dfd31e). Parent ref: group-v275816. {{(pid=63088) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 848.980747] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 848.981150] env[63088]: DEBUG oslo_concurrency.lockutils [None req-31e98785-0c60-4844-bb02-c955550de6ad tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Lock "7f341d80-450a-47b8-b26d-15f2ce3e378c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 57.742s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 848.981360] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cce63fe0-1739-444e-9621-414c6c8f0527 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.991598] env[63088]: INFO nova.virt.vmwareapi.vm_util [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Created folder: Project (dace8b5181b84623b08f903d12dfd31e) in parent group-v275816. [ 848.991778] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Creating folder: Instances. Parent ref: group-v275875. {{(pid=63088) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 848.992051] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c8495b08-e7f8-4d9b-a619-4b98c97501ce {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.001938] env[63088]: INFO nova.virt.vmwareapi.vm_util [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Created folder: Instances in parent group-v275875. [ 849.002096] env[63088]: DEBUG oslo.service.loopingcall [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 849.002302] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Creating VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 849.002511] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c3c063d1-ae7c-4876-9e43-e946251afa5c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.021703] env[63088]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 849.021703] env[63088]: value = "task-1284996" [ 849.021703] env[63088]: _type = "Task" [ 849.021703] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.031779] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c39e0557-36eb-432c-8e51-8136da99476a tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.031779] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1284996, 'name': CreateVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.087665] env[63088]: DEBUG nova.network.neutron [req-a3d6bbb4-2df7-4511-85b6-a75e44638117 req-a88f7a96-f32b-4449-a5de-c888511175e9 service nova] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 849.182021] env[63088]: DEBUG oslo_vmware.api [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1284993, 'name': ReconfigVM_Task, 'duration_secs': 0.956501} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.182932] env[63088]: DEBUG nova.network.neutron [req-a3d6bbb4-2df7-4511-85b6-a75e44638117 req-a88f7a96-f32b-4449-a5de-c888511175e9 service nova] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 849.184199] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 8b564e29-96cf-4abf-963d-142b413fb464] Reconfigured VM instance instance-00000041 to attach disk [datastore1] 8b564e29-96cf-4abf-963d-142b413fb464/8b564e29-96cf-4abf-963d-142b413fb464.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 849.185090] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1384d2e7-8f4f-4781-b606-ba0ce97c3de3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.187135] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f74cf871-55f9-46f3-b9cb-4b34c2963a40 tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Acquiring lock "00925f94-dbf9-453c-a124-d8434679aedf" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.187408] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f74cf871-55f9-46f3-b9cb-4b34c2963a40 tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Lock "00925f94-dbf9-453c-a124-d8434679aedf" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.187645] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f74cf871-55f9-46f3-b9cb-4b34c2963a40 tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Acquiring lock "00925f94-dbf9-453c-a124-d8434679aedf-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.187865] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f74cf871-55f9-46f3-b9cb-4b34c2963a40 tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Lock "00925f94-dbf9-453c-a124-d8434679aedf-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.188079] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f74cf871-55f9-46f3-b9cb-4b34c2963a40 tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Lock "00925f94-dbf9-453c-a124-d8434679aedf-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.190216] env[63088]: INFO nova.compute.manager [None req-f74cf871-55f9-46f3-b9cb-4b34c2963a40 tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 00925f94-dbf9-453c-a124-d8434679aedf] Terminating instance [ 849.192912] env[63088]: DEBUG nova.compute.manager [None req-f74cf871-55f9-46f3-b9cb-4b34c2963a40 tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 00925f94-dbf9-453c-a124-d8434679aedf] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 849.193175] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-f74cf871-55f9-46f3-b9cb-4b34c2963a40 tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 00925f94-dbf9-453c-a124-d8434679aedf] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 849.193980] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecda59f6-6dc0-40e6-9b31-844e2e9fe18f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.198729] env[63088]: DEBUG oslo_vmware.api [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Waiting for the task: (returnval){ [ 849.198729] env[63088]: value = "task-1284997" [ 849.198729] env[63088]: _type = "Task" [ 849.198729] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.204265] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-f74cf871-55f9-46f3-b9cb-4b34c2963a40 tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 00925f94-dbf9-453c-a124-d8434679aedf] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 849.204875] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e6a502b9-26d9-4807-9204-9866a2344d74 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.211493] env[63088]: DEBUG oslo_vmware.api [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1284997, 'name': Rename_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.215840] env[63088]: DEBUG oslo_vmware.api [None req-f74cf871-55f9-46f3-b9cb-4b34c2963a40 tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Waiting for the task: (returnval){ [ 849.215840] env[63088]: value = "task-1284998" [ 849.215840] env[63088]: _type = "Task" [ 849.215840] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.223484] env[63088]: DEBUG oslo_vmware.api [None req-f74cf871-55f9-46f3-b9cb-4b34c2963a40 tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': task-1284998, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.281304] env[63088]: DEBUG oslo_concurrency.lockutils [None req-2790830b-05cd-47fb-bd8c-69ead0f38b2f tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Acquiring lock "7f341d80-450a-47b8-b26d-15f2ce3e378c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.281675] env[63088]: DEBUG oslo_concurrency.lockutils [None req-2790830b-05cd-47fb-bd8c-69ead0f38b2f tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Lock "7f341d80-450a-47b8-b26d-15f2ce3e378c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.281955] env[63088]: DEBUG oslo_concurrency.lockutils [None req-2790830b-05cd-47fb-bd8c-69ead0f38b2f tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Acquiring lock "7f341d80-450a-47b8-b26d-15f2ce3e378c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.282245] env[63088]: DEBUG oslo_concurrency.lockutils [None req-2790830b-05cd-47fb-bd8c-69ead0f38b2f tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Lock "7f341d80-450a-47b8-b26d-15f2ce3e378c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.282449] env[63088]: DEBUG oslo_concurrency.lockutils [None req-2790830b-05cd-47fb-bd8c-69ead0f38b2f tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Lock "7f341d80-450a-47b8-b26d-15f2ce3e378c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.284853] env[63088]: INFO nova.compute.manager [None req-2790830b-05cd-47fb-bd8c-69ead0f38b2f tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 7f341d80-450a-47b8-b26d-15f2ce3e378c] Terminating instance [ 849.286806] env[63088]: DEBUG nova.compute.manager [None req-2790830b-05cd-47fb-bd8c-69ead0f38b2f tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 7f341d80-450a-47b8-b26d-15f2ce3e378c] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 849.287079] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-2790830b-05cd-47fb-bd8c-69ead0f38b2f tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 7f341d80-450a-47b8-b26d-15f2ce3e378c] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 849.287879] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5346ea92-ebb4-431d-bae1-4ba6dd7850c4 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.295534] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-2790830b-05cd-47fb-bd8c-69ead0f38b2f tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 7f341d80-450a-47b8-b26d-15f2ce3e378c] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 849.295721] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-44b6b7de-e931-4af8-aa26-fa4919e576cf {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.298533] env[63088]: DEBUG nova.compute.utils [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 849.302762] env[63088]: DEBUG nova.compute.manager [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 849.302954] env[63088]: DEBUG nova.network.neutron [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 849.310169] env[63088]: DEBUG oslo_vmware.api [None req-2790830b-05cd-47fb-bd8c-69ead0f38b2f tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Waiting for the task: (returnval){ [ 849.310169] env[63088]: value = "task-1284999" [ 849.310169] env[63088]: _type = "Task" [ 849.310169] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.320333] env[63088]: DEBUG oslo_vmware.api [None req-2790830b-05cd-47fb-bd8c-69ead0f38b2f tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': task-1284999, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.368244] env[63088]: DEBUG nova.policy [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9b77ca5186e5437a862d644b2075fe8b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '982309216b714182b5304c4d15808b9e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 849.531809] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1284996, 'name': CreateVM_Task, 'duration_secs': 0.32664} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.531986] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Created VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 849.532676] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 849.532839] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 849.533170] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 849.533422] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2f79ebaa-e099-4589-874c-217372840801 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.539189] env[63088]: DEBUG oslo_vmware.api [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for the task: (returnval){ [ 849.539189] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52973f05-f211-80b8-73ef-b62ac6e98255" [ 849.539189] env[63088]: _type = "Task" [ 849.539189] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.547541] env[63088]: DEBUG oslo_vmware.api [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52973f05-f211-80b8-73ef-b62ac6e98255, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.563157] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b213eed-998a-4d53-9d7a-2e7c52101fb4 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.570219] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ed4aadf-a20e-4502-a423-6a81fe7eedf7 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.602125] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cde6165-fb2a-4e31-889a-2558be8c1995 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.610092] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1c893d3-b17f-4a33-82dc-3574a6366d0a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.624614] env[63088]: DEBUG nova.compute.provider_tree [None req-a26b18b9-d003-4a4b-9d6c-96d60267d634 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 849.645225] env[63088]: DEBUG nova.network.neutron [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Successfully created port: 63d8dbbe-4550-4d8d-8e85-9a9eadc81d82 {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 849.685982] env[63088]: DEBUG oslo_concurrency.lockutils [req-a3d6bbb4-2df7-4511-85b6-a75e44638117 req-a88f7a96-f32b-4449-a5de-c888511175e9 service nova] Releasing lock "refresh_cache-dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 849.686474] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Acquired lock "refresh_cache-dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 849.686665] env[63088]: DEBUG nova.network.neutron [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 849.710576] env[63088]: DEBUG oslo_vmware.api [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1284997, 'name': Rename_Task, 'duration_secs': 0.143729} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.710880] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 8b564e29-96cf-4abf-963d-142b413fb464] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 849.711165] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-18497885-7d48-4ed8-8d65-6ce2fff291f6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.721271] env[63088]: DEBUG oslo_vmware.api [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Waiting for the task: (returnval){ [ 849.721271] env[63088]: value = "task-1285000" [ 849.721271] env[63088]: _type = "Task" [ 849.721271] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.727782] env[63088]: DEBUG oslo_vmware.api [None req-f74cf871-55f9-46f3-b9cb-4b34c2963a40 tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': task-1284998, 'name': PowerOffVM_Task, 'duration_secs': 0.152686} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.728398] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-f74cf871-55f9-46f3-b9cb-4b34c2963a40 tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 00925f94-dbf9-453c-a124-d8434679aedf] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 849.728604] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-f74cf871-55f9-46f3-b9cb-4b34c2963a40 tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 00925f94-dbf9-453c-a124-d8434679aedf] Unregistering the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 849.728853] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-586f70ea-a7e6-4560-8697-68f9547daf1c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.733714] env[63088]: DEBUG oslo_vmware.api [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285000, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.799917] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-f74cf871-55f9-46f3-b9cb-4b34c2963a40 tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 00925f94-dbf9-453c-a124-d8434679aedf] Unregistered the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 849.800183] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-f74cf871-55f9-46f3-b9cb-4b34c2963a40 tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 00925f94-dbf9-453c-a124-d8434679aedf] Deleting contents of the VM from datastore datastore2 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 849.800348] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-f74cf871-55f9-46f3-b9cb-4b34c2963a40 tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Deleting the datastore file [datastore2] 00925f94-dbf9-453c-a124-d8434679aedf {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 849.800628] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-39278f4d-702c-4a67-9c7a-4f255d2f26f4 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.803428] env[63088]: DEBUG nova.compute.manager [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 849.810771] env[63088]: DEBUG oslo_vmware.api [None req-f74cf871-55f9-46f3-b9cb-4b34c2963a40 tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Waiting for the task: (returnval){ [ 849.810771] env[63088]: value = "task-1285002" [ 849.810771] env[63088]: _type = "Task" [ 849.810771] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.822347] env[63088]: DEBUG oslo_vmware.api [None req-f74cf871-55f9-46f3-b9cb-4b34c2963a40 tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': task-1285002, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.825209] env[63088]: DEBUG oslo_vmware.api [None req-2790830b-05cd-47fb-bd8c-69ead0f38b2f tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': task-1284999, 'name': PowerOffVM_Task, 'duration_secs': 0.158125} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.825458] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-2790830b-05cd-47fb-bd8c-69ead0f38b2f tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 7f341d80-450a-47b8-b26d-15f2ce3e378c] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 849.825623] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-2790830b-05cd-47fb-bd8c-69ead0f38b2f tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 7f341d80-450a-47b8-b26d-15f2ce3e378c] Unregistering the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 849.825989] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0fbdb6df-b0f9-4fb4-b657-4168c83bec1a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.880844] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-2790830b-05cd-47fb-bd8c-69ead0f38b2f tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 7f341d80-450a-47b8-b26d-15f2ce3e378c] Unregistered the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 849.881137] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-2790830b-05cd-47fb-bd8c-69ead0f38b2f tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 7f341d80-450a-47b8-b26d-15f2ce3e378c] Deleting contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 849.881345] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-2790830b-05cd-47fb-bd8c-69ead0f38b2f tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Deleting the datastore file [datastore1] 7f341d80-450a-47b8-b26d-15f2ce3e378c {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 849.881983] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-92064be6-a5c8-439c-a774-15f232e0adaa {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.890226] env[63088]: DEBUG oslo_vmware.api [None req-2790830b-05cd-47fb-bd8c-69ead0f38b2f tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Waiting for the task: (returnval){ [ 849.890226] env[63088]: value = "task-1285004" [ 849.890226] env[63088]: _type = "Task" [ 849.890226] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.901454] env[63088]: DEBUG oslo_vmware.api [None req-2790830b-05cd-47fb-bd8c-69ead0f38b2f tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': task-1285004, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.049252] env[63088]: DEBUG oslo_vmware.api [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52973f05-f211-80b8-73ef-b62ac6e98255, 'name': SearchDatastore_Task, 'duration_secs': 0.009846} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.049587] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 850.049822] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Processing image 1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 850.050102] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 850.050262] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.050454] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 850.050738] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-39899009-1de0-47e2-a372-7be9e27c552e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.058667] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 850.058850] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 850.059623] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8a0c77ff-688b-406d-bac1-92303f82a970 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.064585] env[63088]: DEBUG oslo_vmware.api [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for the task: (returnval){ [ 850.064585] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]520574b4-a0a8-a62f-5da9-c028db12cceb" [ 850.064585] env[63088]: _type = "Task" [ 850.064585] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.072031] env[63088]: DEBUG oslo_vmware.api [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]520574b4-a0a8-a62f-5da9-c028db12cceb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.128814] env[63088]: DEBUG nova.scheduler.client.report [None req-a26b18b9-d003-4a4b-9d6c-96d60267d634 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 850.231971] env[63088]: DEBUG oslo_vmware.api [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285000, 'name': PowerOnVM_Task, 'duration_secs': 0.50112} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.232804] env[63088]: DEBUG nova.network.neutron [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 850.234653] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 8b564e29-96cf-4abf-963d-142b413fb464] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 850.236030] env[63088]: INFO nova.compute.manager [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 8b564e29-96cf-4abf-963d-142b413fb464] Took 8.19 seconds to spawn the instance on the hypervisor. [ 850.236030] env[63088]: DEBUG nova.compute.manager [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 8b564e29-96cf-4abf-963d-142b413fb464] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 850.236030] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dafb30b8-dd3f-4b7f-9480-3f3066047a59 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.321623] env[63088]: DEBUG oslo_vmware.api [None req-f74cf871-55f9-46f3-b9cb-4b34c2963a40 tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': task-1285002, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.131754} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.321871] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-f74cf871-55f9-46f3-b9cb-4b34c2963a40 tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 850.322073] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-f74cf871-55f9-46f3-b9cb-4b34c2963a40 tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 00925f94-dbf9-453c-a124-d8434679aedf] Deleted contents of the VM from datastore datastore2 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 850.322257] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-f74cf871-55f9-46f3-b9cb-4b34c2963a40 tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 00925f94-dbf9-453c-a124-d8434679aedf] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 850.322427] env[63088]: INFO nova.compute.manager [None req-f74cf871-55f9-46f3-b9cb-4b34c2963a40 tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 00925f94-dbf9-453c-a124-d8434679aedf] Took 1.13 seconds to destroy the instance on the hypervisor. [ 850.322870] env[63088]: DEBUG oslo.service.loopingcall [None req-f74cf871-55f9-46f3-b9cb-4b34c2963a40 tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 850.322870] env[63088]: DEBUG nova.compute.manager [-] [instance: 00925f94-dbf9-453c-a124-d8434679aedf] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 850.323036] env[63088]: DEBUG nova.network.neutron [-] [instance: 00925f94-dbf9-453c-a124-d8434679aedf] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 850.378016] env[63088]: DEBUG nova.network.neutron [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Updating instance_info_cache with network_info: [{"id": "6ce9925d-cf59-4de0-aece-da90fb4b9f36", "address": "fa:16:3e:b0:ce:19", "network": {"id": "93b4f1f6-4529-4871-87dd-fbbe1fe59ea5", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-559913751-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7ed1e04b27ee464eb9a72866d8327352", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60badc2d-69d2-467d-a92e-98511f5cb0b2", "external-id": "cl2-zone-408", "segmentation_id": 408, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6ce9925d-cf", "ovs_interfaceid": "6ce9925d-cf59-4de0-aece-da90fb4b9f36", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 850.399863] env[63088]: DEBUG oslo_vmware.api [None req-2790830b-05cd-47fb-bd8c-69ead0f38b2f tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': task-1285004, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.165873} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.400135] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-2790830b-05cd-47fb-bd8c-69ead0f38b2f tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 850.400317] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-2790830b-05cd-47fb-bd8c-69ead0f38b2f tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 7f341d80-450a-47b8-b26d-15f2ce3e378c] Deleted contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 850.400491] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-2790830b-05cd-47fb-bd8c-69ead0f38b2f tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 7f341d80-450a-47b8-b26d-15f2ce3e378c] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 850.400655] env[63088]: INFO nova.compute.manager [None req-2790830b-05cd-47fb-bd8c-69ead0f38b2f tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 7f341d80-450a-47b8-b26d-15f2ce3e378c] Took 1.11 seconds to destroy the instance on the hypervisor. [ 850.400892] env[63088]: DEBUG oslo.service.loopingcall [None req-2790830b-05cd-47fb-bd8c-69ead0f38b2f tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 850.401097] env[63088]: DEBUG nova.compute.manager [-] [instance: 7f341d80-450a-47b8-b26d-15f2ce3e378c] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 850.401163] env[63088]: DEBUG nova.network.neutron [-] [instance: 7f341d80-450a-47b8-b26d-15f2ce3e378c] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 850.581499] env[63088]: DEBUG oslo_vmware.api [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]520574b4-a0a8-a62f-5da9-c028db12cceb, 'name': SearchDatastore_Task, 'duration_secs': 0.007643} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.582340] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cd1092c4-0c2a-4034-ac51-7bb5f1c15f32 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.589989] env[63088]: DEBUG oslo_vmware.api [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for the task: (returnval){ [ 850.589989] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52d374e4-c276-9d24-90c4-3a4464d775f9" [ 850.589989] env[63088]: _type = "Task" [ 850.589989] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.592991] env[63088]: DEBUG nova.compute.manager [req-9431652e-3547-44ce-a6fc-32d912d881b0 req-dea002f7-5c4d-4030-975a-97c21d6fd75a service nova] [instance: 00925f94-dbf9-453c-a124-d8434679aedf] Received event network-vif-deleted-e2779ab6-aee9-4245-aab0-9d598a612844 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 850.593194] env[63088]: INFO nova.compute.manager [req-9431652e-3547-44ce-a6fc-32d912d881b0 req-dea002f7-5c4d-4030-975a-97c21d6fd75a service nova] [instance: 00925f94-dbf9-453c-a124-d8434679aedf] Neutron deleted interface e2779ab6-aee9-4245-aab0-9d598a612844; detaching it from the instance and deleting it from the info cache [ 850.593367] env[63088]: DEBUG nova.network.neutron [req-9431652e-3547-44ce-a6fc-32d912d881b0 req-dea002f7-5c4d-4030-975a-97c21d6fd75a service nova] [instance: 00925f94-dbf9-453c-a124-d8434679aedf] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 850.603029] env[63088]: DEBUG oslo_vmware.api [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52d374e4-c276-9d24-90c4-3a4464d775f9, 'name': SearchDatastore_Task, 'duration_secs': 0.009104} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.604097] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 850.604097] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] 7b6aadb7-e34b-42b7-b69f-370434f5b665/7b6aadb7-e34b-42b7-b69f-370434f5b665.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 850.604097] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4f347aac-e8e4-43c6-871d-870494411161 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.611216] env[63088]: DEBUG oslo_vmware.api [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for the task: (returnval){ [ 850.611216] env[63088]: value = "task-1285005" [ 850.611216] env[63088]: _type = "Task" [ 850.611216] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.620077] env[63088]: DEBUG oslo_vmware.api [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285005, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.631968] env[63088]: DEBUG nova.compute.manager [req-d1a3db5c-27e4-4e28-8add-717c4f9aa44b req-5e958ec4-177e-4afa-b3e3-4a900f3f3fd3 service nova] [instance: 7f341d80-450a-47b8-b26d-15f2ce3e378c] Received event network-vif-deleted-711f2731-680e-4328-97c8-0a2d1bbf79eb {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 850.632164] env[63088]: INFO nova.compute.manager [req-d1a3db5c-27e4-4e28-8add-717c4f9aa44b req-5e958ec4-177e-4afa-b3e3-4a900f3f3fd3 service nova] [instance: 7f341d80-450a-47b8-b26d-15f2ce3e378c] Neutron deleted interface 711f2731-680e-4328-97c8-0a2d1bbf79eb; detaching it from the instance and deleting it from the info cache [ 850.632743] env[63088]: DEBUG nova.network.neutron [req-d1a3db5c-27e4-4e28-8add-717c4f9aa44b req-5e958ec4-177e-4afa-b3e3-4a900f3f3fd3 service nova] [instance: 7f341d80-450a-47b8-b26d-15f2ce3e378c] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 850.635042] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a26b18b9-d003-4a4b-9d6c-96d60267d634 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.839s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.636736] env[63088]: DEBUG oslo_concurrency.lockutils [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.371s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 850.638145] env[63088]: INFO nova.compute.claims [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: 58d561fe-26aa-4e94-8d55-cc70c361b479] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 850.656527] env[63088]: INFO nova.scheduler.client.report [None req-a26b18b9-d003-4a4b-9d6c-96d60267d634 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Deleted allocations for instance 4f6e2d99-7300-42e6-8abf-362ddd4e8c6d [ 850.756051] env[63088]: INFO nova.compute.manager [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 8b564e29-96cf-4abf-963d-142b413fb464] Took 22.76 seconds to build instance. [ 850.815191] env[63088]: DEBUG nova.compute.manager [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 850.844462] env[63088]: DEBUG nova.virt.hardware [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:18:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='99d18326-0562-4fb9-afc4-e9e0d5c92bb4',id=38,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-1781853035',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 850.844705] env[63088]: DEBUG nova.virt.hardware [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 850.844863] env[63088]: DEBUG nova.virt.hardware [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 850.845054] env[63088]: DEBUG nova.virt.hardware [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 850.845204] env[63088]: DEBUG nova.virt.hardware [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 850.845361] env[63088]: DEBUG nova.virt.hardware [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 850.845565] env[63088]: DEBUG nova.virt.hardware [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 850.845703] env[63088]: DEBUG nova.virt.hardware [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 850.845866] env[63088]: DEBUG nova.virt.hardware [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 850.846042] env[63088]: DEBUG nova.virt.hardware [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 850.846218] env[63088]: DEBUG nova.virt.hardware [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 850.847118] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-203037e5-f2af-440b-a104-bf193ec7a41e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.855219] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d1232bc-a445-4f42-98f6-9d96e6d048d3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.881234] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Releasing lock "refresh_cache-dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 850.881667] env[63088]: DEBUG nova.compute.manager [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Instance network_info: |[{"id": "6ce9925d-cf59-4de0-aece-da90fb4b9f36", "address": "fa:16:3e:b0:ce:19", "network": {"id": "93b4f1f6-4529-4871-87dd-fbbe1fe59ea5", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-559913751-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7ed1e04b27ee464eb9a72866d8327352", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60badc2d-69d2-467d-a92e-98511f5cb0b2", "external-id": "cl2-zone-408", "segmentation_id": 408, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6ce9925d-cf", "ovs_interfaceid": "6ce9925d-cf59-4de0-aece-da90fb4b9f36", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 850.882340] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b0:ce:19', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '60badc2d-69d2-467d-a92e-98511f5cb0b2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6ce9925d-cf59-4de0-aece-da90fb4b9f36', 'vif_model': 'vmxnet3'}] {{(pid=63088) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 850.893653] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Creating folder: Project (7ed1e04b27ee464eb9a72866d8327352). Parent ref: group-v275816. {{(pid=63088) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 850.894089] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4768441f-ebea-45e5-adb1-4c7a25fc41fb {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.904939] env[63088]: INFO nova.virt.vmwareapi.vm_util [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Created folder: Project (7ed1e04b27ee464eb9a72866d8327352) in parent group-v275816. [ 850.905181] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Creating folder: Instances. Parent ref: group-v275878. {{(pid=63088) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 850.905435] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-67ddd112-4552-4558-a7ff-85af1bce9bce {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.914885] env[63088]: INFO nova.virt.vmwareapi.vm_util [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Created folder: Instances in parent group-v275878. [ 850.915169] env[63088]: DEBUG oslo.service.loopingcall [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 850.915380] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Creating VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 850.915595] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5bd43d35-ea1b-4780-90ff-decfa340bbd7 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.936458] env[63088]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 850.936458] env[63088]: value = "task-1285008" [ 850.936458] env[63088]: _type = "Task" [ 850.936458] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.946386] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285008, 'name': CreateVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.029443] env[63088]: DEBUG nova.network.neutron [-] [instance: 00925f94-dbf9-453c-a124-d8434679aedf] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 851.101028] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-527ad144-79f9-40ca-9e24-2146163f5fea {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.112353] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d44c0e7-3598-43e3-9639-1686636a3538 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.124651] env[63088]: DEBUG nova.network.neutron [-] [instance: 7f341d80-450a-47b8-b26d-15f2ce3e378c] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 851.135428] env[63088]: DEBUG oslo_vmware.api [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285005, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.457782} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.135428] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] 7b6aadb7-e34b-42b7-b69f-370434f5b665/7b6aadb7-e34b-42b7-b69f-370434f5b665.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 851.135648] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Extending root virtual disk to 1048576 {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 851.135681] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4ce52444-bf37-4f6f-b731-5dc1270630a6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.137956] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c664e5ab-13d6-47e5-afa5-6d0cf85494b2 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.151748] env[63088]: DEBUG nova.compute.manager [req-9431652e-3547-44ce-a6fc-32d912d881b0 req-dea002f7-5c4d-4030-975a-97c21d6fd75a service nova] [instance: 00925f94-dbf9-453c-a124-d8434679aedf] Detach interface failed, port_id=e2779ab6-aee9-4245-aab0-9d598a612844, reason: Instance 00925f94-dbf9-453c-a124-d8434679aedf could not be found. {{(pid=63088) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 851.156264] env[63088]: DEBUG oslo_vmware.api [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for the task: (returnval){ [ 851.156264] env[63088]: value = "task-1285009" [ 851.156264] env[63088]: _type = "Task" [ 851.156264] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.171890] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9bd702d-f86b-4b22-93dd-1d5c54955557 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.182648] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a26b18b9-d003-4a4b-9d6c-96d60267d634 tempest-InstanceActionsV221TestJSON-1284712409 tempest-InstanceActionsV221TestJSON-1284712409-project-member] Lock "4f6e2d99-7300-42e6-8abf-362ddd4e8c6d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.745s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.187747] env[63088]: DEBUG oslo_vmware.api [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285009, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.205944] env[63088]: DEBUG nova.compute.manager [req-d1a3db5c-27e4-4e28-8add-717c4f9aa44b req-5e958ec4-177e-4afa-b3e3-4a900f3f3fd3 service nova] [instance: 7f341d80-450a-47b8-b26d-15f2ce3e378c] Detach interface failed, port_id=711f2731-680e-4328-97c8-0a2d1bbf79eb, reason: Instance 7f341d80-450a-47b8-b26d-15f2ce3e378c could not be found. {{(pid=63088) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 851.260623] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c7297251-bb70-4ab2-9fb1-58f62820e885 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Lock "8b564e29-96cf-4abf-963d-142b413fb464" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 59.164s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.447769] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285008, 'name': CreateVM_Task, 'duration_secs': 0.350197} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.448011] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Created VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 851.448691] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 851.448852] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.449194] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 851.449466] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-564c620f-1f43-46ed-9055-6cb9223c2d05 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.454131] env[63088]: DEBUG oslo_vmware.api [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Waiting for the task: (returnval){ [ 851.454131] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]528ed08d-b2bc-2f0e-889b-9816c82aa416" [ 851.454131] env[63088]: _type = "Task" [ 851.454131] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.463670] env[63088]: DEBUG oslo_vmware.api [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]528ed08d-b2bc-2f0e-889b-9816c82aa416, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.533031] env[63088]: INFO nova.compute.manager [-] [instance: 00925f94-dbf9-453c-a124-d8434679aedf] Took 1.21 seconds to deallocate network for instance. [ 851.627883] env[63088]: INFO nova.compute.manager [-] [instance: 7f341d80-450a-47b8-b26d-15f2ce3e378c] Took 1.23 seconds to deallocate network for instance. [ 851.665258] env[63088]: DEBUG oslo_vmware.api [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285009, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074574} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.665543] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Extended root virtual disk {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 851.666387] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53b0eeb1-4e62-49b5-8182-8de1d31957f5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.691365] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Reconfiguring VM instance instance-00000042 to attach disk [datastore1] 7b6aadb7-e34b-42b7-b69f-370434f5b665/7b6aadb7-e34b-42b7-b69f-370434f5b665.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 851.694886] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7c296656-31d9-4e9a-a8bb-c2157b09f18a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.716648] env[63088]: DEBUG oslo_vmware.api [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for the task: (returnval){ [ 851.716648] env[63088]: value = "task-1285010" [ 851.716648] env[63088]: _type = "Task" [ 851.716648] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.725567] env[63088]: DEBUG oslo_vmware.api [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285010, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.772415] env[63088]: DEBUG nova.network.neutron [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Successfully updated port: 63d8dbbe-4550-4d8d-8e85-9a9eadc81d82 {{(pid=63088) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 851.963315] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac7d4f95-86c5-4be3-a3d4-c9bd9936da10 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.968456] env[63088]: DEBUG oslo_vmware.api [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]528ed08d-b2bc-2f0e-889b-9816c82aa416, 'name': SearchDatastore_Task, 'duration_secs': 0.009142} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.969136] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 851.969375] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Processing image 1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 851.969601] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 851.970087] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.970087] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 851.970506] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5df42fc0-4810-4402-9c5d-b24e99831231 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.974531] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58184f2c-89a7-4cfb-b05f-afa91be27932 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.979219] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 851.979302] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 852.005185] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-382c5745-9717-483c-be0c-839bdbbed0a2 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.008230] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83c67161-9b10-4af6-83b4-892919b4cf6d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.013543] env[63088]: DEBUG oslo_vmware.api [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Waiting for the task: (returnval){ [ 852.013543] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52449e63-6b5a-0837-9367-8d88bc1c42d8" [ 852.013543] env[63088]: _type = "Task" [ 852.013543] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.019191] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56b064b5-0a9b-49c4-bc4d-3d5676c7b546 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.028147] env[63088]: DEBUG oslo_vmware.api [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52449e63-6b5a-0837-9367-8d88bc1c42d8, 'name': SearchDatastore_Task, 'duration_secs': 0.008307} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.037127] env[63088]: DEBUG nova.compute.provider_tree [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 852.038349] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6ea5ef2e-4aa7-4c70-ba91-3847b95f9930 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.041869] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f74cf871-55f9-46f3-b9cb-4b34c2963a40 tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 852.045746] env[63088]: DEBUG oslo_vmware.api [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Waiting for the task: (returnval){ [ 852.045746] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52265d84-a2f5-d31f-0538-4649b319713e" [ 852.045746] env[63088]: _type = "Task" [ 852.045746] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.055026] env[63088]: DEBUG oslo_vmware.api [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52265d84-a2f5-d31f-0538-4649b319713e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.135259] env[63088]: DEBUG oslo_concurrency.lockutils [None req-2790830b-05cd-47fb-bd8c-69ead0f38b2f tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 852.226298] env[63088]: DEBUG oslo_vmware.api [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285010, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.279810] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Acquiring lock "refresh_cache-e6b0ce08-d67f-458f-92f5-1e904d03a5ef" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 852.279810] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Acquired lock "refresh_cache-e6b0ce08-d67f-458f-92f5-1e904d03a5ef" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 852.279810] env[63088]: DEBUG nova.network.neutron [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 852.290747] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad69f81b-3386-4ace-9568-18242911b544 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.297949] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-01054098-bc8e-4ff7-a635-6e6fe72aac68 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 8b564e29-96cf-4abf-963d-142b413fb464] Suspending the VM {{(pid=63088) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1163}} [ 852.298573] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-da4c3dc2-9617-4e82-8459-aceefa139811 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.306815] env[63088]: DEBUG oslo_vmware.api [None req-01054098-bc8e-4ff7-a635-6e6fe72aac68 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Waiting for the task: (returnval){ [ 852.306815] env[63088]: value = "task-1285011" [ 852.306815] env[63088]: _type = "Task" [ 852.306815] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.319113] env[63088]: DEBUG oslo_vmware.api [None req-01054098-bc8e-4ff7-a635-6e6fe72aac68 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285011, 'name': SuspendVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.544282] env[63088]: DEBUG nova.scheduler.client.report [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 852.563558] env[63088]: DEBUG oslo_vmware.api [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52265d84-a2f5-d31f-0538-4649b319713e, 'name': SearchDatastore_Task, 'duration_secs': 0.009068} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.564430] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.564723] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore2] dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69/dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 852.566388] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-04071e65-de5b-4c2c-b83c-a89d235d02ab {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.573816] env[63088]: DEBUG oslo_vmware.api [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Waiting for the task: (returnval){ [ 852.573816] env[63088]: value = "task-1285012" [ 852.573816] env[63088]: _type = "Task" [ 852.573816] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.583623] env[63088]: DEBUG oslo_vmware.api [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Task: {'id': task-1285012, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.621499] env[63088]: DEBUG nova.compute.manager [req-167805c5-9bdc-4a35-8984-929fc984a026 req-5593c1d1-00e7-4f27-b942-7c315d836e27 service nova] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Received event network-vif-plugged-63d8dbbe-4550-4d8d-8e85-9a9eadc81d82 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 852.621793] env[63088]: DEBUG oslo_concurrency.lockutils [req-167805c5-9bdc-4a35-8984-929fc984a026 req-5593c1d1-00e7-4f27-b942-7c315d836e27 service nova] Acquiring lock "e6b0ce08-d67f-458f-92f5-1e904d03a5ef-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 852.622414] env[63088]: DEBUG oslo_concurrency.lockutils [req-167805c5-9bdc-4a35-8984-929fc984a026 req-5593c1d1-00e7-4f27-b942-7c315d836e27 service nova] Lock "e6b0ce08-d67f-458f-92f5-1e904d03a5ef-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 852.622723] env[63088]: DEBUG oslo_concurrency.lockutils [req-167805c5-9bdc-4a35-8984-929fc984a026 req-5593c1d1-00e7-4f27-b942-7c315d836e27 service nova] Lock "e6b0ce08-d67f-458f-92f5-1e904d03a5ef-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.623086] env[63088]: DEBUG nova.compute.manager [req-167805c5-9bdc-4a35-8984-929fc984a026 req-5593c1d1-00e7-4f27-b942-7c315d836e27 service nova] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] No waiting events found dispatching network-vif-plugged-63d8dbbe-4550-4d8d-8e85-9a9eadc81d82 {{(pid=63088) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 852.623397] env[63088]: WARNING nova.compute.manager [req-167805c5-9bdc-4a35-8984-929fc984a026 req-5593c1d1-00e7-4f27-b942-7c315d836e27 service nova] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Received unexpected event network-vif-plugged-63d8dbbe-4550-4d8d-8e85-9a9eadc81d82 for instance with vm_state building and task_state spawning. [ 852.623686] env[63088]: DEBUG nova.compute.manager [req-167805c5-9bdc-4a35-8984-929fc984a026 req-5593c1d1-00e7-4f27-b942-7c315d836e27 service nova] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Received event network-changed-63d8dbbe-4550-4d8d-8e85-9a9eadc81d82 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 852.623957] env[63088]: DEBUG nova.compute.manager [req-167805c5-9bdc-4a35-8984-929fc984a026 req-5593c1d1-00e7-4f27-b942-7c315d836e27 service nova] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Refreshing instance network info cache due to event network-changed-63d8dbbe-4550-4d8d-8e85-9a9eadc81d82. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 852.624251] env[63088]: DEBUG oslo_concurrency.lockutils [req-167805c5-9bdc-4a35-8984-929fc984a026 req-5593c1d1-00e7-4f27-b942-7c315d836e27 service nova] Acquiring lock "refresh_cache-e6b0ce08-d67f-458f-92f5-1e904d03a5ef" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 852.728729] env[63088]: DEBUG oslo_vmware.api [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285010, 'name': ReconfigVM_Task, 'duration_secs': 0.600508} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.728993] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Reconfigured VM instance instance-00000042 to attach disk [datastore1] 7b6aadb7-e34b-42b7-b69f-370434f5b665/7b6aadb7-e34b-42b7-b69f-370434f5b665.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 852.729699] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-487305fd-03c6-45d8-85cb-a9477e018487 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.736865] env[63088]: DEBUG oslo_vmware.api [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for the task: (returnval){ [ 852.736865] env[63088]: value = "task-1285013" [ 852.736865] env[63088]: _type = "Task" [ 852.736865] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.747176] env[63088]: DEBUG oslo_vmware.api [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285013, 'name': Rename_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.816826] env[63088]: DEBUG oslo_vmware.api [None req-01054098-bc8e-4ff7-a635-6e6fe72aac68 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285011, 'name': SuspendVM_Task} progress is 50%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.828405] env[63088]: DEBUG nova.network.neutron [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 853.056494] env[63088]: DEBUG nova.network.neutron [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Updating instance_info_cache with network_info: [{"id": "63d8dbbe-4550-4d8d-8e85-9a9eadc81d82", "address": "fa:16:3e:d1:72:56", "network": {"id": "ebeefbc2-043c-4e76-aeaf-4f9113869855", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.60", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "bfb6bc5062fc4d51af1d3c577659be9a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271fe7a0-dfd7-409b-920a-cf83ef1a86a3", "external-id": "nsx-vlan-transportzone-728", "segmentation_id": 728, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap63d8dbbe-45", "ovs_interfaceid": "63d8dbbe-4550-4d8d-8e85-9a9eadc81d82", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 853.058297] env[63088]: DEBUG oslo_concurrency.lockutils [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.421s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.058734] env[63088]: DEBUG nova.compute.manager [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: 58d561fe-26aa-4e94-8d55-cc70c361b479] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 853.061328] env[63088]: DEBUG oslo_concurrency.lockutils [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 17.055s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 853.061507] env[63088]: DEBUG nova.objects.instance [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63088) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 853.085571] env[63088]: DEBUG oslo_vmware.api [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Task: {'id': task-1285012, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.247731] env[63088]: DEBUG oslo_vmware.api [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285013, 'name': Rename_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.318450] env[63088]: DEBUG oslo_vmware.api [None req-01054098-bc8e-4ff7-a635-6e6fe72aac68 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285011, 'name': SuspendVM_Task} progress is 100%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.562411] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Releasing lock "refresh_cache-e6b0ce08-d67f-458f-92f5-1e904d03a5ef" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 853.562845] env[63088]: DEBUG nova.compute.manager [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Instance network_info: |[{"id": "63d8dbbe-4550-4d8d-8e85-9a9eadc81d82", "address": "fa:16:3e:d1:72:56", "network": {"id": "ebeefbc2-043c-4e76-aeaf-4f9113869855", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.60", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "bfb6bc5062fc4d51af1d3c577659be9a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271fe7a0-dfd7-409b-920a-cf83ef1a86a3", "external-id": "nsx-vlan-transportzone-728", "segmentation_id": 728, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap63d8dbbe-45", "ovs_interfaceid": "63d8dbbe-4550-4d8d-8e85-9a9eadc81d82", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 853.563784] env[63088]: DEBUG oslo_concurrency.lockutils [req-167805c5-9bdc-4a35-8984-929fc984a026 req-5593c1d1-00e7-4f27-b942-7c315d836e27 service nova] Acquired lock "refresh_cache-e6b0ce08-d67f-458f-92f5-1e904d03a5ef" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 853.563784] env[63088]: DEBUG nova.network.neutron [req-167805c5-9bdc-4a35-8984-929fc984a026 req-5593c1d1-00e7-4f27-b942-7c315d836e27 service nova] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Refreshing network info cache for port 63d8dbbe-4550-4d8d-8e85-9a9eadc81d82 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 853.564927] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d1:72:56', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '271fe7a0-dfd7-409b-920a-cf83ef1a86a3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '63d8dbbe-4550-4d8d-8e85-9a9eadc81d82', 'vif_model': 'vmxnet3'}] {{(pid=63088) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 853.572219] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Creating folder: Project (982309216b714182b5304c4d15808b9e). Parent ref: group-v275816. {{(pid=63088) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 853.579554] env[63088]: DEBUG nova.compute.utils [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 853.580867] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-daaa1051-e5cc-401a-8ed5-af873c03b3c4 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.583894] env[63088]: DEBUG nova.compute.manager [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: 58d561fe-26aa-4e94-8d55-cc70c361b479] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 853.583894] env[63088]: DEBUG nova.network.neutron [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: 58d561fe-26aa-4e94-8d55-cc70c361b479] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 853.595617] env[63088]: DEBUG oslo_vmware.api [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Task: {'id': task-1285012, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.598706] env[63088]: INFO nova.virt.vmwareapi.vm_util [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Created folder: Project (982309216b714182b5304c4d15808b9e) in parent group-v275816. [ 853.598984] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Creating folder: Instances. Parent ref: group-v275881. {{(pid=63088) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 853.599252] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-022fd138-11e5-4013-8ae1-f73828fbf149 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.609286] env[63088]: INFO nova.virt.vmwareapi.vm_util [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Created folder: Instances in parent group-v275881. [ 853.609547] env[63088]: DEBUG oslo.service.loopingcall [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 853.609809] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Creating VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 853.610147] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-74fd0099-e34c-4833-b884-b2e3d0bdb527 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.635335] env[63088]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 853.635335] env[63088]: value = "task-1285016" [ 853.635335] env[63088]: _type = "Task" [ 853.635335] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.644568] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285016, 'name': CreateVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.653149] env[63088]: DEBUG nova.policy [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '98509947cc6745ef8037ff68da8925d2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '900466c3c09a4f928bdd4b602715c838', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 853.747720] env[63088]: DEBUG oslo_vmware.api [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285013, 'name': Rename_Task, 'duration_secs': 0.68478} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.748041] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 853.748318] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9279336a-880d-41da-b1aa-7201c7faf2f7 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.754890] env[63088]: DEBUG oslo_vmware.api [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for the task: (returnval){ [ 853.754890] env[63088]: value = "task-1285017" [ 853.754890] env[63088]: _type = "Task" [ 853.754890] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.764807] env[63088]: DEBUG oslo_vmware.api [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285017, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.817320] env[63088]: DEBUG oslo_vmware.api [None req-01054098-bc8e-4ff7-a635-6e6fe72aac68 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285011, 'name': SuspendVM_Task, 'duration_secs': 1.048336} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.817589] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-01054098-bc8e-4ff7-a635-6e6fe72aac68 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 8b564e29-96cf-4abf-963d-142b413fb464] Suspended the VM {{(pid=63088) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1167}} [ 853.817954] env[63088]: DEBUG nova.compute.manager [None req-01054098-bc8e-4ff7-a635-6e6fe72aac68 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 8b564e29-96cf-4abf-963d-142b413fb464] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 853.818983] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f087a540-5802-4941-8d15-a6aaac88e473 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.831117] env[63088]: DEBUG nova.network.neutron [req-167805c5-9bdc-4a35-8984-929fc984a026 req-5593c1d1-00e7-4f27-b942-7c315d836e27 service nova] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Updated VIF entry in instance network info cache for port 63d8dbbe-4550-4d8d-8e85-9a9eadc81d82. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 853.831432] env[63088]: DEBUG nova.network.neutron [req-167805c5-9bdc-4a35-8984-929fc984a026 req-5593c1d1-00e7-4f27-b942-7c315d836e27 service nova] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Updating instance_info_cache with network_info: [{"id": "63d8dbbe-4550-4d8d-8e85-9a9eadc81d82", "address": "fa:16:3e:d1:72:56", "network": {"id": "ebeefbc2-043c-4e76-aeaf-4f9113869855", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.60", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "bfb6bc5062fc4d51af1d3c577659be9a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271fe7a0-dfd7-409b-920a-cf83ef1a86a3", "external-id": "nsx-vlan-transportzone-728", "segmentation_id": 728, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap63d8dbbe-45", "ovs_interfaceid": "63d8dbbe-4550-4d8d-8e85-9a9eadc81d82", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 854.049780] env[63088]: DEBUG nova.network.neutron [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: 58d561fe-26aa-4e94-8d55-cc70c361b479] Successfully created port: c6477c64-3129-4546-84b1-40bc79bcaa57 {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 854.084009] env[63088]: DEBUG nova.compute.manager [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: 58d561fe-26aa-4e94-8d55-cc70c361b479] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 854.087476] env[63088]: DEBUG oslo_concurrency.lockutils [None req-10a82445-e0a0-4916-a64c-440e82e18079 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.026s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.092967] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.794s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.094284] env[63088]: INFO nova.compute.claims [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 9d5f1cde-e787-4a2c-ac49-83ec135ff51c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 854.102772] env[63088]: DEBUG oslo_vmware.api [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Task: {'id': task-1285012, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.161711} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.103159] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore2] dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69/dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 854.103636] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Extending root virtual disk to 1048576 {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 854.104015] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-aba3521c-4071-4115-8da7-c993dd66eeb1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.110866] env[63088]: DEBUG oslo_vmware.api [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Waiting for the task: (returnval){ [ 854.110866] env[63088]: value = "task-1285018" [ 854.110866] env[63088]: _type = "Task" [ 854.110866] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.119831] env[63088]: DEBUG oslo_vmware.api [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Task: {'id': task-1285018, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.151146] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285016, 'name': CreateVM_Task, 'duration_secs': 0.373526} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.154016] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Created VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 854.155505] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 854.155505] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.155863] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 854.157079] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-276db0d9-7fbf-4b6d-a8d9-73e7733df7a7 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.164035] env[63088]: DEBUG oslo_vmware.api [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Waiting for the task: (returnval){ [ 854.164035] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]5259cda2-57a8-12a7-ca7a-9218cc8dbb0f" [ 854.164035] env[63088]: _type = "Task" [ 854.164035] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.173389] env[63088]: DEBUG oslo_vmware.api [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]5259cda2-57a8-12a7-ca7a-9218cc8dbb0f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.266777] env[63088]: DEBUG oslo_vmware.api [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285017, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.334146] env[63088]: DEBUG oslo_concurrency.lockutils [req-167805c5-9bdc-4a35-8984-929fc984a026 req-5593c1d1-00e7-4f27-b942-7c315d836e27 service nova] Releasing lock "refresh_cache-e6b0ce08-d67f-458f-92f5-1e904d03a5ef" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 854.625413] env[63088]: DEBUG oslo_vmware.api [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Task: {'id': task-1285018, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06919} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.625702] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Extended root virtual disk {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 854.626643] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a5cbff3-3556-4d34-b748-df40199e33e7 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.653682] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Reconfiguring VM instance instance-00000043 to attach disk [datastore2] dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69/dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 854.654231] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0b59b304-4eed-4ea7-9237-d67a96512643 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.687167] env[63088]: DEBUG oslo_vmware.api [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]5259cda2-57a8-12a7-ca7a-9218cc8dbb0f, 'name': SearchDatastore_Task, 'duration_secs': 0.020931} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.689593] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 854.689593] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Processing image 1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 854.689593] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 854.689593] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.689844] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 854.690035] env[63088]: DEBUG oslo_vmware.api [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Waiting for the task: (returnval){ [ 854.690035] env[63088]: value = "task-1285019" [ 854.690035] env[63088]: _type = "Task" [ 854.690035] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.690255] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bb37f3e3-e749-4da9-a649-6aff713ba6e9 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.705483] env[63088]: DEBUG oslo_vmware.api [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Task: {'id': task-1285019, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.708018] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 854.708263] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 854.709052] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ccfa4339-fe59-4d8a-8348-a62d15edf1aa {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.715367] env[63088]: DEBUG oslo_vmware.api [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Waiting for the task: (returnval){ [ 854.715367] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]5270fc44-c34d-ec3d-e5d4-bf53008eab7d" [ 854.715367] env[63088]: _type = "Task" [ 854.715367] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.727574] env[63088]: DEBUG oslo_vmware.api [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]5270fc44-c34d-ec3d-e5d4-bf53008eab7d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.766269] env[63088]: DEBUG oslo_vmware.api [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285017, 'name': PowerOnVM_Task, 'duration_secs': 0.527114} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.766520] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 854.766720] env[63088]: INFO nova.compute.manager [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Took 10.18 seconds to spawn the instance on the hypervisor. [ 854.766899] env[63088]: DEBUG nova.compute.manager [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 854.767750] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fa5a9ee-9e51-4c66-b07b-0c1367ed6853 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.102182] env[63088]: DEBUG nova.compute.manager [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: 58d561fe-26aa-4e94-8d55-cc70c361b479] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 855.133738] env[63088]: DEBUG nova.virt.hardware [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 855.133987] env[63088]: DEBUG nova.virt.hardware [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 855.134163] env[63088]: DEBUG nova.virt.hardware [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 855.134361] env[63088]: DEBUG nova.virt.hardware [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 855.134541] env[63088]: DEBUG nova.virt.hardware [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 855.134659] env[63088]: DEBUG nova.virt.hardware [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 855.134868] env[63088]: DEBUG nova.virt.hardware [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 855.135040] env[63088]: DEBUG nova.virt.hardware [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 855.135215] env[63088]: DEBUG nova.virt.hardware [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 855.135374] env[63088]: DEBUG nova.virt.hardware [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 855.135614] env[63088]: DEBUG nova.virt.hardware [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 855.136841] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c604079-3949-484e-9b95-eaf511080eaf {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.144647] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-841fc86c-d139-42d4-b4b5-3e71826ef440 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.209474] env[63088]: DEBUG oslo_vmware.api [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Task: {'id': task-1285019, 'name': ReconfigVM_Task, 'duration_secs': 0.302596} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.213658] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Reconfigured VM instance instance-00000043 to attach disk [datastore2] dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69/dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 855.213658] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4d3e684d-8637-4673-8b65-ee0449dfa4fa {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.220825] env[63088]: DEBUG oslo_vmware.api [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Waiting for the task: (returnval){ [ 855.220825] env[63088]: value = "task-1285020" [ 855.220825] env[63088]: _type = "Task" [ 855.220825] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.231416] env[63088]: DEBUG oslo_vmware.api [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]5270fc44-c34d-ec3d-e5d4-bf53008eab7d, 'name': SearchDatastore_Task, 'duration_secs': 0.01476} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.232632] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b6b808fc-f14a-456d-b598-c7c8d523791b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.242521] env[63088]: DEBUG oslo_vmware.api [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Task: {'id': task-1285020, 'name': Rename_Task} progress is 10%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.249329] env[63088]: DEBUG oslo_vmware.api [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Waiting for the task: (returnval){ [ 855.249329] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]526ae68f-5c58-aaaa-c333-ded508bc6d22" [ 855.249329] env[63088]: _type = "Task" [ 855.249329] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.258160] env[63088]: DEBUG oslo_vmware.api [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]526ae68f-5c58-aaaa-c333-ded508bc6d22, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.295536] env[63088]: INFO nova.compute.manager [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Took 27.19 seconds to build instance. [ 855.421362] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d66c8fbc-3034-4760-a672-3962e4163616 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.429690] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4002ecd-e18c-49ce-aaa6-2d276dd98e4e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.462707] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6c3e100-0ae6-4abe-8318-28025ccf8a4e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.469979] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73f99617-f8f4-4a4e-bfca-8c6526fbedd8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.483031] env[63088]: DEBUG nova.compute.provider_tree [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 855.562119] env[63088]: DEBUG nova.compute.manager [req-afcd2cb7-e57c-43dc-ad57-edeadc830dc9 req-380a169f-df81-4b8c-81f9-f05d8713e830 service nova] [instance: 58d561fe-26aa-4e94-8d55-cc70c361b479] Received event network-vif-plugged-c6477c64-3129-4546-84b1-40bc79bcaa57 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 855.562119] env[63088]: DEBUG oslo_concurrency.lockutils [req-afcd2cb7-e57c-43dc-ad57-edeadc830dc9 req-380a169f-df81-4b8c-81f9-f05d8713e830 service nova] Acquiring lock "58d561fe-26aa-4e94-8d55-cc70c361b479-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 855.562119] env[63088]: DEBUG oslo_concurrency.lockutils [req-afcd2cb7-e57c-43dc-ad57-edeadc830dc9 req-380a169f-df81-4b8c-81f9-f05d8713e830 service nova] Lock "58d561fe-26aa-4e94-8d55-cc70c361b479-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 855.562119] env[63088]: DEBUG oslo_concurrency.lockutils [req-afcd2cb7-e57c-43dc-ad57-edeadc830dc9 req-380a169f-df81-4b8c-81f9-f05d8713e830 service nova] Lock "58d561fe-26aa-4e94-8d55-cc70c361b479-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 855.562119] env[63088]: DEBUG nova.compute.manager [req-afcd2cb7-e57c-43dc-ad57-edeadc830dc9 req-380a169f-df81-4b8c-81f9-f05d8713e830 service nova] [instance: 58d561fe-26aa-4e94-8d55-cc70c361b479] No waiting events found dispatching network-vif-plugged-c6477c64-3129-4546-84b1-40bc79bcaa57 {{(pid=63088) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 855.562279] env[63088]: WARNING nova.compute.manager [req-afcd2cb7-e57c-43dc-ad57-edeadc830dc9 req-380a169f-df81-4b8c-81f9-f05d8713e830 service nova] [instance: 58d561fe-26aa-4e94-8d55-cc70c361b479] Received unexpected event network-vif-plugged-c6477c64-3129-4546-84b1-40bc79bcaa57 for instance with vm_state building and task_state spawning. [ 855.643226] env[63088]: DEBUG nova.network.neutron [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: 58d561fe-26aa-4e94-8d55-cc70c361b479] Successfully updated port: c6477c64-3129-4546-84b1-40bc79bcaa57 {{(pid=63088) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 855.686948] env[63088]: DEBUG nova.compute.manager [None req-0bd62baa-37b1-44d3-beee-8ebfe632ec9b tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 8b564e29-96cf-4abf-963d-142b413fb464] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 855.686948] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58398468-5c6a-4d9c-96a0-dfd5ae364c6e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.731256] env[63088]: DEBUG oslo_vmware.api [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Task: {'id': task-1285020, 'name': Rename_Task} progress is 99%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.759698] env[63088]: DEBUG oslo_vmware.api [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]526ae68f-5c58-aaaa-c333-ded508bc6d22, 'name': SearchDatastore_Task, 'duration_secs': 0.010821} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.759986] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 855.760442] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore2] e6b0ce08-d67f-458f-92f5-1e904d03a5ef/e6b0ce08-d67f-458f-92f5-1e904d03a5ef.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 855.761134] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5ab8b53f-229e-4406-9ff1-df07cad452c7 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.767610] env[63088]: DEBUG oslo_vmware.api [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Waiting for the task: (returnval){ [ 855.767610] env[63088]: value = "task-1285021" [ 855.767610] env[63088]: _type = "Task" [ 855.767610] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.776282] env[63088]: DEBUG oslo_vmware.api [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Task: {'id': task-1285021, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.797982] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a83fb09a-c044-47dc-b60e-713abc27ab29 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lock "7b6aadb7-e34b-42b7-b69f-370434f5b665" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 58.566s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 855.989022] env[63088]: DEBUG nova.scheduler.client.report [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 856.147590] env[63088]: DEBUG oslo_concurrency.lockutils [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Acquiring lock "refresh_cache-58d561fe-26aa-4e94-8d55-cc70c361b479" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 856.147590] env[63088]: DEBUG oslo_concurrency.lockutils [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Acquired lock "refresh_cache-58d561fe-26aa-4e94-8d55-cc70c361b479" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 856.147590] env[63088]: DEBUG nova.network.neutron [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: 58d561fe-26aa-4e94-8d55-cc70c361b479] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 856.197601] env[63088]: INFO nova.compute.manager [None req-0bd62baa-37b1-44d3-beee-8ebfe632ec9b tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 8b564e29-96cf-4abf-963d-142b413fb464] instance snapshotting [ 856.198460] env[63088]: WARNING nova.compute.manager [None req-0bd62baa-37b1-44d3-beee-8ebfe632ec9b tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 8b564e29-96cf-4abf-963d-142b413fb464] trying to snapshot a non-running instance: (state: 7 expected: 1) [ 856.201472] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74dcf536-b931-4487-befa-484e63ae1958 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.227267] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d962543-6a53-414a-910e-53c381e99551 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.236270] env[63088]: DEBUG oslo_vmware.api [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Task: {'id': task-1285020, 'name': Rename_Task} progress is 99%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.281546] env[63088]: DEBUG oslo_vmware.api [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Task: {'id': task-1285021, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.492033] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.399s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.492477] env[63088]: DEBUG nova.compute.manager [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 9d5f1cde-e787-4a2c-ac49-83ec135ff51c] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 856.495250] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0d443006-c001-488d-9c7f-f5cbe25ea71a tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.258s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 856.495471] env[63088]: DEBUG nova.objects.instance [None req-0d443006-c001-488d-9c7f-f5cbe25ea71a tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Lazy-loading 'resources' on Instance uuid 4847f15e-98d4-401c-91ff-067e84a85727 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 856.681514] env[63088]: DEBUG nova.network.neutron [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: 58d561fe-26aa-4e94-8d55-cc70c361b479] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 856.734035] env[63088]: DEBUG oslo_vmware.api [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Task: {'id': task-1285020, 'name': Rename_Task, 'duration_secs': 1.146707} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.734478] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 856.737132] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4dd5be2a-23c9-4532-8dbc-94a846129c6c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.743648] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-0bd62baa-37b1-44d3-beee-8ebfe632ec9b tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 8b564e29-96cf-4abf-963d-142b413fb464] Creating Snapshot of the VM instance {{(pid=63088) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 856.745159] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-9a4691d0-10c1-460d-b7fe-6dac6058a9cb {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.747145] env[63088]: DEBUG oslo_vmware.api [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Waiting for the task: (returnval){ [ 856.747145] env[63088]: value = "task-1285022" [ 856.747145] env[63088]: _type = "Task" [ 856.747145] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.755211] env[63088]: DEBUG oslo_vmware.api [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Task: {'id': task-1285022, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.756641] env[63088]: DEBUG oslo_vmware.api [None req-0bd62baa-37b1-44d3-beee-8ebfe632ec9b tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Waiting for the task: (returnval){ [ 856.756641] env[63088]: value = "task-1285023" [ 856.756641] env[63088]: _type = "Task" [ 856.756641] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.764793] env[63088]: DEBUG oslo_vmware.api [None req-0bd62baa-37b1-44d3-beee-8ebfe632ec9b tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285023, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.780431] env[63088]: DEBUG oslo_vmware.api [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Task: {'id': task-1285021, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.647} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.780431] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore2] e6b0ce08-d67f-458f-92f5-1e904d03a5ef/e6b0ce08-d67f-458f-92f5-1e904d03a5ef.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 856.780627] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Extending root virtual disk to 1048576 {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 856.780729] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0ad17116-46b2-4182-a736-a9ff6a605676 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.787352] env[63088]: DEBUG oslo_vmware.api [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Waiting for the task: (returnval){ [ 856.787352] env[63088]: value = "task-1285024" [ 856.787352] env[63088]: _type = "Task" [ 856.787352] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.795873] env[63088]: DEBUG oslo_vmware.api [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Task: {'id': task-1285024, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.895570] env[63088]: DEBUG nova.network.neutron [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: 58d561fe-26aa-4e94-8d55-cc70c361b479] Updating instance_info_cache with network_info: [{"id": "c6477c64-3129-4546-84b1-40bc79bcaa57", "address": "fa:16:3e:75:e7:1a", "network": {"id": "7db9302e-1c7a-42dc-8084-98e2a57ad7bf", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-2073161169-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "900466c3c09a4f928bdd4b602715c838", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "87bbf4e0-9064-4516-b7e7-44973f817205", "external-id": "nsx-vlan-transportzone-507", "segmentation_id": 507, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6477c64-31", "ovs_interfaceid": "c6477c64-3129-4546-84b1-40bc79bcaa57", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 857.002426] env[63088]: DEBUG nova.compute.utils [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 857.012036] env[63088]: DEBUG nova.compute.manager [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 9d5f1cde-e787-4a2c-ac49-83ec135ff51c] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 857.012036] env[63088]: DEBUG nova.network.neutron [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 9d5f1cde-e787-4a2c-ac49-83ec135ff51c] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 857.065077] env[63088]: DEBUG nova.policy [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '139b65d9cf314b3db1d685d63bc19bba', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e48f62dee9ad4e9b94b67bd871db5c63', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 857.267351] env[63088]: DEBUG oslo_vmware.api [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Task: {'id': task-1285022, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.273698] env[63088]: DEBUG oslo_vmware.api [None req-0bd62baa-37b1-44d3-beee-8ebfe632ec9b tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285023, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.297676] env[63088]: DEBUG oslo_vmware.api [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Task: {'id': task-1285024, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.057028} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.300853] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Extended root virtual disk {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 857.305650] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-574e089a-fcfa-4782-9d43-545cddc0bdbe {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.311639] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8aa4b42-03db-4153-b0ff-1f871a60a967 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.333753] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Reconfiguring VM instance instance-00000044 to attach disk [datastore2] e6b0ce08-d67f-458f-92f5-1e904d03a5ef/e6b0ce08-d67f-458f-92f5-1e904d03a5ef.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 857.334664] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b54abe2a-a11e-4598-bd49-b5635daff9f3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.353919] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1121d562-e6ea-44b1-9832-eecab979b7ad {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.359669] env[63088]: DEBUG oslo_vmware.api [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Waiting for the task: (returnval){ [ 857.359669] env[63088]: value = "task-1285025" [ 857.359669] env[63088]: _type = "Task" [ 857.359669] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.389654] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d6afe06-af9f-4cdd-8d65-b95b9608803b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.393183] env[63088]: DEBUG nova.network.neutron [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 9d5f1cde-e787-4a2c-ac49-83ec135ff51c] Successfully created port: 07e1b344-3f50-473d-a443-caca3ba6c0ea {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 857.398073] env[63088]: DEBUG oslo_concurrency.lockutils [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Releasing lock "refresh_cache-58d561fe-26aa-4e94-8d55-cc70c361b479" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 857.398361] env[63088]: DEBUG nova.compute.manager [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: 58d561fe-26aa-4e94-8d55-cc70c361b479] Instance network_info: |[{"id": "c6477c64-3129-4546-84b1-40bc79bcaa57", "address": "fa:16:3e:75:e7:1a", "network": {"id": "7db9302e-1c7a-42dc-8084-98e2a57ad7bf", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-2073161169-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "900466c3c09a4f928bdd4b602715c838", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "87bbf4e0-9064-4516-b7e7-44973f817205", "external-id": "nsx-vlan-transportzone-507", "segmentation_id": 507, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6477c64-31", "ovs_interfaceid": "c6477c64-3129-4546-84b1-40bc79bcaa57", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 857.398645] env[63088]: DEBUG oslo_vmware.api [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Task: {'id': task-1285025, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.399599] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: 58d561fe-26aa-4e94-8d55-cc70c361b479] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:75:e7:1a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '87bbf4e0-9064-4516-b7e7-44973f817205', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c6477c64-3129-4546-84b1-40bc79bcaa57', 'vif_model': 'vmxnet3'}] {{(pid=63088) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 857.406999] env[63088]: DEBUG oslo.service.loopingcall [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 857.409212] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 58d561fe-26aa-4e94-8d55-cc70c361b479] Creating VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 857.409489] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b2f14645-6668-4848-a8ed-48f7388b8c22 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.430034] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cc4daab-17ea-4433-af28-f5bda9de1048 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.440452] env[63088]: DEBUG nova.compute.provider_tree [None req-0d443006-c001-488d-9c7f-f5cbe25ea71a tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 857.444014] env[63088]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 857.444014] env[63088]: value = "task-1285026" [ 857.444014] env[63088]: _type = "Task" [ 857.444014] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.451625] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285026, 'name': CreateVM_Task} progress is 6%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.513407] env[63088]: DEBUG nova.compute.manager [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 9d5f1cde-e787-4a2c-ac49-83ec135ff51c] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 857.639892] env[63088]: DEBUG nova.compute.manager [req-e6cca2d3-fae7-4ae8-8bb3-7eea7611e9da req-49365ffa-faa9-48b8-9cf5-a5898c986254 service nova] [instance: 58d561fe-26aa-4e94-8d55-cc70c361b479] Received event network-changed-c6477c64-3129-4546-84b1-40bc79bcaa57 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 857.640676] env[63088]: DEBUG nova.compute.manager [req-e6cca2d3-fae7-4ae8-8bb3-7eea7611e9da req-49365ffa-faa9-48b8-9cf5-a5898c986254 service nova] [instance: 58d561fe-26aa-4e94-8d55-cc70c361b479] Refreshing instance network info cache due to event network-changed-c6477c64-3129-4546-84b1-40bc79bcaa57. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 857.640676] env[63088]: DEBUG oslo_concurrency.lockutils [req-e6cca2d3-fae7-4ae8-8bb3-7eea7611e9da req-49365ffa-faa9-48b8-9cf5-a5898c986254 service nova] Acquiring lock "refresh_cache-58d561fe-26aa-4e94-8d55-cc70c361b479" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 857.641044] env[63088]: DEBUG oslo_concurrency.lockutils [req-e6cca2d3-fae7-4ae8-8bb3-7eea7611e9da req-49365ffa-faa9-48b8-9cf5-a5898c986254 service nova] Acquired lock "refresh_cache-58d561fe-26aa-4e94-8d55-cc70c361b479" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 857.641292] env[63088]: DEBUG nova.network.neutron [req-e6cca2d3-fae7-4ae8-8bb3-7eea7611e9da req-49365ffa-faa9-48b8-9cf5-a5898c986254 service nova] [instance: 58d561fe-26aa-4e94-8d55-cc70c361b479] Refreshing network info cache for port c6477c64-3129-4546-84b1-40bc79bcaa57 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 857.647627] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Acquiring lock "667fd5e9-5fe4-41e1-9d8a-896c6e5c6286" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 857.647627] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Lock "667fd5e9-5fe4-41e1-9d8a-896c6e5c6286" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 857.679156] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Acquiring lock "b50db882-598c-488f-b935-34f55f655642" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 857.679597] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Lock "b50db882-598c-488f-b935-34f55f655642" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 857.708506] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Acquiring lock "2964db35-7357-40a7-b4e6-7e2595549f5b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 857.709036] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Lock "2964db35-7357-40a7-b4e6-7e2595549f5b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 857.759976] env[63088]: DEBUG oslo_vmware.api [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Task: {'id': task-1285022, 'name': PowerOnVM_Task, 'duration_secs': 0.524493} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.763877] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 857.764185] env[63088]: INFO nova.compute.manager [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Took 10.79 seconds to spawn the instance on the hypervisor. [ 857.764428] env[63088]: DEBUG nova.compute.manager [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 857.765247] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afb612b3-ab4c-4b6a-a487-d23e59de6597 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.776771] env[63088]: DEBUG oslo_vmware.api [None req-0bd62baa-37b1-44d3-beee-8ebfe632ec9b tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285023, 'name': CreateSnapshot_Task, 'duration_secs': 0.605615} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.780808] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-0bd62baa-37b1-44d3-beee-8ebfe632ec9b tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 8b564e29-96cf-4abf-963d-142b413fb464] Created Snapshot of the VM instance {{(pid=63088) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 857.781782] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15be2746-1428-41f6-9caf-6b6addbe02e3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.870337] env[63088]: DEBUG oslo_vmware.api [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Task: {'id': task-1285025, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.948034] env[63088]: DEBUG nova.scheduler.client.report [None req-0d443006-c001-488d-9c7f-f5cbe25ea71a tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 857.959694] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285026, 'name': CreateVM_Task} progress is 99%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.150208] env[63088]: DEBUG nova.compute.manager [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: 667fd5e9-5fe4-41e1-9d8a-896c6e5c6286] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 858.183872] env[63088]: DEBUG nova.compute.manager [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: b50db882-598c-488f-b935-34f55f655642] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 858.211398] env[63088]: DEBUG nova.compute.manager [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: 2964db35-7357-40a7-b4e6-7e2595549f5b] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 858.290118] env[63088]: INFO nova.compute.manager [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Took 28.23 seconds to build instance. [ 858.302075] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-0bd62baa-37b1-44d3-beee-8ebfe632ec9b tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 8b564e29-96cf-4abf-963d-142b413fb464] Creating linked-clone VM from snapshot {{(pid=63088) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 858.302386] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-ee0b6e90-055a-40f9-9325-02e5b031e6d3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.311647] env[63088]: DEBUG oslo_vmware.api [None req-0bd62baa-37b1-44d3-beee-8ebfe632ec9b tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Waiting for the task: (returnval){ [ 858.311647] env[63088]: value = "task-1285027" [ 858.311647] env[63088]: _type = "Task" [ 858.311647] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.319852] env[63088]: DEBUG oslo_vmware.api [None req-0bd62baa-37b1-44d3-beee-8ebfe632ec9b tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285027, 'name': CloneVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.356575] env[63088]: DEBUG nova.network.neutron [req-e6cca2d3-fae7-4ae8-8bb3-7eea7611e9da req-49365ffa-faa9-48b8-9cf5-a5898c986254 service nova] [instance: 58d561fe-26aa-4e94-8d55-cc70c361b479] Updated VIF entry in instance network info cache for port c6477c64-3129-4546-84b1-40bc79bcaa57. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 858.356934] env[63088]: DEBUG nova.network.neutron [req-e6cca2d3-fae7-4ae8-8bb3-7eea7611e9da req-49365ffa-faa9-48b8-9cf5-a5898c986254 service nova] [instance: 58d561fe-26aa-4e94-8d55-cc70c361b479] Updating instance_info_cache with network_info: [{"id": "c6477c64-3129-4546-84b1-40bc79bcaa57", "address": "fa:16:3e:75:e7:1a", "network": {"id": "7db9302e-1c7a-42dc-8084-98e2a57ad7bf", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-2073161169-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "900466c3c09a4f928bdd4b602715c838", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "87bbf4e0-9064-4516-b7e7-44973f817205", "external-id": "nsx-vlan-transportzone-507", "segmentation_id": 507, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6477c64-31", "ovs_interfaceid": "c6477c64-3129-4546-84b1-40bc79bcaa57", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 858.369380] env[63088]: DEBUG oslo_vmware.api [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Task: {'id': task-1285025, 'name': ReconfigVM_Task, 'duration_secs': 0.686962} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.370231] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Reconfigured VM instance instance-00000044 to attach disk [datastore2] e6b0ce08-d67f-458f-92f5-1e904d03a5ef/e6b0ce08-d67f-458f-92f5-1e904d03a5ef.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 858.370887] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ccc38968-e819-4462-b4b9-b6cac6aa5e28 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.378020] env[63088]: DEBUG oslo_vmware.api [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Waiting for the task: (returnval){ [ 858.378020] env[63088]: value = "task-1285028" [ 858.378020] env[63088]: _type = "Task" [ 858.378020] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.385774] env[63088]: DEBUG oslo_vmware.api [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Task: {'id': task-1285028, 'name': Rename_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.455186] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0d443006-c001-488d-9c7f-f5cbe25ea71a tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.960s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.462122] env[63088]: DEBUG oslo_concurrency.lockutils [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 17.111s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.462122] env[63088]: DEBUG nova.objects.instance [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] [instance: b159649b-ebd5-4c7d-9074-dc0e6395a947] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63088) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 858.464793] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285026, 'name': CreateVM_Task, 'duration_secs': 0.526572} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.465318] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 58d561fe-26aa-4e94-8d55-cc70c361b479] Created VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 858.469176] env[63088]: DEBUG oslo_concurrency.lockutils [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 858.469176] env[63088]: DEBUG oslo_concurrency.lockutils [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.469176] env[63088]: DEBUG oslo_concurrency.lockutils [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 858.469176] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0b8b5316-5a96-453c-b255-a1053be2eb24 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.472259] env[63088]: DEBUG oslo_vmware.api [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Waiting for the task: (returnval){ [ 858.472259] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52017dd5-d4a7-9336-fa9a-3fbd38781d0a" [ 858.472259] env[63088]: _type = "Task" [ 858.472259] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.483371] env[63088]: DEBUG oslo_vmware.api [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52017dd5-d4a7-9336-fa9a-3fbd38781d0a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.489029] env[63088]: INFO nova.scheduler.client.report [None req-0d443006-c001-488d-9c7f-f5cbe25ea71a tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Deleted allocations for instance 4847f15e-98d4-401c-91ff-067e84a85727 [ 858.522981] env[63088]: DEBUG nova.compute.manager [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 9d5f1cde-e787-4a2c-ac49-83ec135ff51c] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 858.548872] env[63088]: DEBUG nova.virt.hardware [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 858.549146] env[63088]: DEBUG nova.virt.hardware [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 858.549368] env[63088]: DEBUG nova.virt.hardware [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 858.549580] env[63088]: DEBUG nova.virt.hardware [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 858.550014] env[63088]: DEBUG nova.virt.hardware [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 858.550014] env[63088]: DEBUG nova.virt.hardware [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 858.550330] env[63088]: DEBUG nova.virt.hardware [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 858.550601] env[63088]: DEBUG nova.virt.hardware [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 858.550815] env[63088]: DEBUG nova.virt.hardware [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 858.551026] env[63088]: DEBUG nova.virt.hardware [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 858.551239] env[63088]: DEBUG nova.virt.hardware [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 858.552518] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8fac32b-1378-471f-b3bc-cc22708e003b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.560585] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9eadb300-02e0-4e34-8fa2-ea52f4bf5581 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.646209] env[63088]: DEBUG oslo_concurrency.lockutils [None req-10c6f8ec-c5b4-4d2b-853d-2d940a3f0da2 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquiring lock "interface-4971b24c-6710-4f50-9846-727dad264b1f-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.646209] env[63088]: DEBUG oslo_concurrency.lockutils [None req-10c6f8ec-c5b4-4d2b-853d-2d940a3f0da2 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Lock "interface-4971b24c-6710-4f50-9846-727dad264b1f-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.646209] env[63088]: DEBUG nova.objects.instance [None req-10c6f8ec-c5b4-4d2b-853d-2d940a3f0da2 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Lazy-loading 'flavor' on Instance uuid 4971b24c-6710-4f50-9846-727dad264b1f {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 858.671456] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.707563] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.737113] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.791951] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4bf0f2b4-3522-4dd4-a826-07a73c19912f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Lock "dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 61.191s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.823404] env[63088]: DEBUG oslo_vmware.api [None req-0bd62baa-37b1-44d3-beee-8ebfe632ec9b tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285027, 'name': CloneVM_Task} progress is 94%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.860621] env[63088]: DEBUG oslo_concurrency.lockutils [req-e6cca2d3-fae7-4ae8-8bb3-7eea7611e9da req-49365ffa-faa9-48b8-9cf5-a5898c986254 service nova] Releasing lock "refresh_cache-58d561fe-26aa-4e94-8d55-cc70c361b479" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 858.861088] env[63088]: DEBUG nova.compute.manager [req-e6cca2d3-fae7-4ae8-8bb3-7eea7611e9da req-49365ffa-faa9-48b8-9cf5-a5898c986254 service nova] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Received event network-changed-fa4a91d1-df9a-4789-bc5c-a8b95457cd93 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 858.861298] env[63088]: DEBUG nova.compute.manager [req-e6cca2d3-fae7-4ae8-8bb3-7eea7611e9da req-49365ffa-faa9-48b8-9cf5-a5898c986254 service nova] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Refreshing instance network info cache due to event network-changed-fa4a91d1-df9a-4789-bc5c-a8b95457cd93. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 858.861522] env[63088]: DEBUG oslo_concurrency.lockutils [req-e6cca2d3-fae7-4ae8-8bb3-7eea7611e9da req-49365ffa-faa9-48b8-9cf5-a5898c986254 service nova] Acquiring lock "refresh_cache-7b6aadb7-e34b-42b7-b69f-370434f5b665" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 858.861667] env[63088]: DEBUG oslo_concurrency.lockutils [req-e6cca2d3-fae7-4ae8-8bb3-7eea7611e9da req-49365ffa-faa9-48b8-9cf5-a5898c986254 service nova] Acquired lock "refresh_cache-7b6aadb7-e34b-42b7-b69f-370434f5b665" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.861831] env[63088]: DEBUG nova.network.neutron [req-e6cca2d3-fae7-4ae8-8bb3-7eea7611e9da req-49365ffa-faa9-48b8-9cf5-a5898c986254 service nova] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Refreshing network info cache for port fa4a91d1-df9a-4789-bc5c-a8b95457cd93 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 858.888638] env[63088]: DEBUG oslo_vmware.api [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Task: {'id': task-1285028, 'name': Rename_Task, 'duration_secs': 0.234929} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.889047] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 858.889374] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a68e2045-ea28-4947-a857-ed58adc52183 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.895841] env[63088]: DEBUG oslo_vmware.api [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Waiting for the task: (returnval){ [ 858.895841] env[63088]: value = "task-1285029" [ 858.895841] env[63088]: _type = "Task" [ 858.895841] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.904355] env[63088]: DEBUG oslo_vmware.api [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Task: {'id': task-1285029, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.983117] env[63088]: DEBUG oslo_vmware.api [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52017dd5-d4a7-9336-fa9a-3fbd38781d0a, 'name': SearchDatastore_Task, 'duration_secs': 0.012281} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.983442] env[63088]: DEBUG oslo_concurrency.lockutils [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 858.983692] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: 58d561fe-26aa-4e94-8d55-cc70c361b479] Processing image 1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 858.983923] env[63088]: DEBUG oslo_concurrency.lockutils [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 858.984080] env[63088]: DEBUG oslo_concurrency.lockutils [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.984259] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 858.984515] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-98da0e49-17e7-4680-b416-9bdff76e09a6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.000022] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 859.000022] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 859.000022] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0d443006-c001-488d-9c7f-f5cbe25ea71a tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Lock "4847f15e-98d4-401c-91ff-067e84a85727" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.150s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 859.000022] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9daf1067-a5b5-44b8-9a9d-2064c66e9482 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.005383] env[63088]: DEBUG oslo_vmware.api [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Waiting for the task: (returnval){ [ 859.005383] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]525f1519-1dd7-810c-f981-dd0d821c2406" [ 859.005383] env[63088]: _type = "Task" [ 859.005383] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.013717] env[63088]: DEBUG oslo_vmware.api [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]525f1519-1dd7-810c-f981-dd0d821c2406, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.150327] env[63088]: DEBUG nova.objects.instance [None req-10c6f8ec-c5b4-4d2b-853d-2d940a3f0da2 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Lazy-loading 'pci_requests' on Instance uuid 4971b24c-6710-4f50-9846-727dad264b1f {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 859.325682] env[63088]: DEBUG oslo_vmware.api [None req-0bd62baa-37b1-44d3-beee-8ebfe632ec9b tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285027, 'name': CloneVM_Task} progress is 94%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.408811] env[63088]: DEBUG oslo_vmware.api [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Task: {'id': task-1285029, 'name': PowerOnVM_Task} progress is 64%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.471558] env[63088]: DEBUG oslo_concurrency.lockutils [None req-66f54f53-3cfd-46d0-b026-cfa276f43d33 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.009s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 859.472988] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.530s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 859.475023] env[63088]: INFO nova.compute.claims [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 859.518388] env[63088]: DEBUG oslo_vmware.api [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]525f1519-1dd7-810c-f981-dd0d821c2406, 'name': SearchDatastore_Task, 'duration_secs': 0.009773} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.521755] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6052eb2d-2904-46bf-a4a9-07c5619b3000 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.527642] env[63088]: DEBUG oslo_vmware.api [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Waiting for the task: (returnval){ [ 859.527642] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52f4d4bf-40f6-b3ec-e7a0-af648cf8554d" [ 859.527642] env[63088]: _type = "Task" [ 859.527642] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.533270] env[63088]: DEBUG nova.network.neutron [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 9d5f1cde-e787-4a2c-ac49-83ec135ff51c] Successfully updated port: 07e1b344-3f50-473d-a443-caca3ba6c0ea {{(pid=63088) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 859.541524] env[63088]: DEBUG oslo_vmware.api [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52f4d4bf-40f6-b3ec-e7a0-af648cf8554d, 'name': SearchDatastore_Task, 'duration_secs': 0.009843} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.541524] env[63088]: DEBUG oslo_concurrency.lockutils [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 859.541936] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore2] 58d561fe-26aa-4e94-8d55-cc70c361b479/58d561fe-26aa-4e94-8d55-cc70c361b479.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 859.542513] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0d9b6517-b2f7-4ccb-b5f7-2ef3dec76b63 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.552466] env[63088]: DEBUG oslo_vmware.api [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Waiting for the task: (returnval){ [ 859.552466] env[63088]: value = "task-1285030" [ 859.552466] env[63088]: _type = "Task" [ 859.552466] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.560699] env[63088]: DEBUG oslo_vmware.api [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': task-1285030, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.653327] env[63088]: DEBUG nova.objects.base [None req-10c6f8ec-c5b4-4d2b-853d-2d940a3f0da2 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Object Instance<4971b24c-6710-4f50-9846-727dad264b1f> lazy-loaded attributes: flavor,pci_requests {{(pid=63088) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 859.653600] env[63088]: DEBUG nova.network.neutron [None req-10c6f8ec-c5b4-4d2b-853d-2d940a3f0da2 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 859.658461] env[63088]: DEBUG nova.network.neutron [req-e6cca2d3-fae7-4ae8-8bb3-7eea7611e9da req-49365ffa-faa9-48b8-9cf5-a5898c986254 service nova] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Updated VIF entry in instance network info cache for port fa4a91d1-df9a-4789-bc5c-a8b95457cd93. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 859.658827] env[63088]: DEBUG nova.network.neutron [req-e6cca2d3-fae7-4ae8-8bb3-7eea7611e9da req-49365ffa-faa9-48b8-9cf5-a5898c986254 service nova] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Updating instance_info_cache with network_info: [{"id": "fa4a91d1-df9a-4789-bc5c-a8b95457cd93", "address": "fa:16:3e:7e:c2:c7", "network": {"id": "dfcbe153-bc01-4362-9247-de9f0b1c847b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-841757397-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.152", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dace8b5181b84623b08f903d12dfd31e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c68b7663-4f0e-47f0-ac7f-40c6d952f7bb", "external-id": "nsx-vlan-transportzone-696", "segmentation_id": 696, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfa4a91d1-df", "ovs_interfaceid": "fa4a91d1-df9a-4789-bc5c-a8b95457cd93", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 859.772371] env[63088]: DEBUG oslo_concurrency.lockutils [None req-10c6f8ec-c5b4-4d2b-853d-2d940a3f0da2 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Lock "interface-4971b24c-6710-4f50-9846-727dad264b1f-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.127s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 859.831024] env[63088]: DEBUG oslo_vmware.api [None req-0bd62baa-37b1-44d3-beee-8ebfe632ec9b tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285027, 'name': CloneVM_Task, 'duration_secs': 1.394809} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.831024] env[63088]: INFO nova.virt.vmwareapi.vmops [None req-0bd62baa-37b1-44d3-beee-8ebfe632ec9b tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 8b564e29-96cf-4abf-963d-142b413fb464] Created linked-clone VM from snapshot [ 859.831024] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e1206c2-99b5-42f3-9c52-134220d57f30 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.838993] env[63088]: DEBUG nova.virt.vmwareapi.images [None req-0bd62baa-37b1-44d3-beee-8ebfe632ec9b tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 8b564e29-96cf-4abf-963d-142b413fb464] Uploading image f1a9d8b0-fef1-49b9-8a87-99a7556efe13 {{(pid=63088) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 859.864883] env[63088]: DEBUG oslo_vmware.rw_handles [None req-0bd62baa-37b1-44d3-beee-8ebfe632ec9b tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 859.864883] env[63088]: value = "vm-275886" [ 859.864883] env[63088]: _type = "VirtualMachine" [ 859.864883] env[63088]: }. {{(pid=63088) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 859.865124] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-9b2fb322-c9e2-45d6-8daf-52f1a4f89fef {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.876176] env[63088]: DEBUG oslo_vmware.rw_handles [None req-0bd62baa-37b1-44d3-beee-8ebfe632ec9b tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Lease: (returnval){ [ 859.876176] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52b3497e-2cf0-920e-cbb4-6599aee947ce" [ 859.876176] env[63088]: _type = "HttpNfcLease" [ 859.876176] env[63088]: } obtained for exporting VM: (result){ [ 859.876176] env[63088]: value = "vm-275886" [ 859.876176] env[63088]: _type = "VirtualMachine" [ 859.876176] env[63088]: }. {{(pid=63088) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 859.876526] env[63088]: DEBUG oslo_vmware.api [None req-0bd62baa-37b1-44d3-beee-8ebfe632ec9b tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Waiting for the lease: (returnval){ [ 859.876526] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52b3497e-2cf0-920e-cbb4-6599aee947ce" [ 859.876526] env[63088]: _type = "HttpNfcLease" [ 859.876526] env[63088]: } to be ready. {{(pid=63088) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 859.887739] env[63088]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 859.887739] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52b3497e-2cf0-920e-cbb4-6599aee947ce" [ 859.887739] env[63088]: _type = "HttpNfcLease" [ 859.887739] env[63088]: } is ready. {{(pid=63088) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 859.888179] env[63088]: DEBUG oslo_vmware.rw_handles [None req-0bd62baa-37b1-44d3-beee-8ebfe632ec9b tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 859.888179] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52b3497e-2cf0-920e-cbb4-6599aee947ce" [ 859.888179] env[63088]: _type = "HttpNfcLease" [ 859.888179] env[63088]: }. {{(pid=63088) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 859.889085] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8662349-0dc6-43b6-89f7-05f3322a1b1a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.899790] env[63088]: DEBUG oslo_vmware.rw_handles [None req-0bd62baa-37b1-44d3-beee-8ebfe632ec9b tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/520bd55d-9d5e-31c0-7922-68afdf00ff83/disk-0.vmdk from lease info. {{(pid=63088) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 859.900132] env[63088]: DEBUG oslo_vmware.rw_handles [None req-0bd62baa-37b1-44d3-beee-8ebfe632ec9b tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/520bd55d-9d5e-31c0-7922-68afdf00ff83/disk-0.vmdk for reading. {{(pid=63088) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 859.975315] env[63088]: DEBUG nova.compute.manager [req-1e08c496-cab0-462f-9375-6266db0c5230 req-baad82cf-2a90-4736-ae3c-df06e3976d11 service nova] [instance: 9d5f1cde-e787-4a2c-ac49-83ec135ff51c] Received event network-vif-plugged-07e1b344-3f50-473d-a443-caca3ba6c0ea {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 859.975543] env[63088]: DEBUG oslo_concurrency.lockutils [req-1e08c496-cab0-462f-9375-6266db0c5230 req-baad82cf-2a90-4736-ae3c-df06e3976d11 service nova] Acquiring lock "9d5f1cde-e787-4a2c-ac49-83ec135ff51c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 859.975743] env[63088]: DEBUG oslo_concurrency.lockutils [req-1e08c496-cab0-462f-9375-6266db0c5230 req-baad82cf-2a90-4736-ae3c-df06e3976d11 service nova] Lock "9d5f1cde-e787-4a2c-ac49-83ec135ff51c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 859.975910] env[63088]: DEBUG oslo_concurrency.lockutils [req-1e08c496-cab0-462f-9375-6266db0c5230 req-baad82cf-2a90-4736-ae3c-df06e3976d11 service nova] Lock "9d5f1cde-e787-4a2c-ac49-83ec135ff51c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 859.976087] env[63088]: DEBUG nova.compute.manager [req-1e08c496-cab0-462f-9375-6266db0c5230 req-baad82cf-2a90-4736-ae3c-df06e3976d11 service nova] [instance: 9d5f1cde-e787-4a2c-ac49-83ec135ff51c] No waiting events found dispatching network-vif-plugged-07e1b344-3f50-473d-a443-caca3ba6c0ea {{(pid=63088) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 859.976253] env[63088]: WARNING nova.compute.manager [req-1e08c496-cab0-462f-9375-6266db0c5230 req-baad82cf-2a90-4736-ae3c-df06e3976d11 service nova] [instance: 9d5f1cde-e787-4a2c-ac49-83ec135ff51c] Received unexpected event network-vif-plugged-07e1b344-3f50-473d-a443-caca3ba6c0ea for instance with vm_state building and task_state spawning. [ 859.976413] env[63088]: DEBUG nova.compute.manager [req-1e08c496-cab0-462f-9375-6266db0c5230 req-baad82cf-2a90-4736-ae3c-df06e3976d11 service nova] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Received event network-changed-6ce9925d-cf59-4de0-aece-da90fb4b9f36 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 859.976562] env[63088]: DEBUG nova.compute.manager [req-1e08c496-cab0-462f-9375-6266db0c5230 req-baad82cf-2a90-4736-ae3c-df06e3976d11 service nova] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Refreshing instance network info cache due to event network-changed-6ce9925d-cf59-4de0-aece-da90fb4b9f36. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 859.976737] env[63088]: DEBUG oslo_concurrency.lockutils [req-1e08c496-cab0-462f-9375-6266db0c5230 req-baad82cf-2a90-4736-ae3c-df06e3976d11 service nova] Acquiring lock "refresh_cache-dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 859.976876] env[63088]: DEBUG oslo_concurrency.lockutils [req-1e08c496-cab0-462f-9375-6266db0c5230 req-baad82cf-2a90-4736-ae3c-df06e3976d11 service nova] Acquired lock "refresh_cache-dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.977040] env[63088]: DEBUG nova.network.neutron [req-1e08c496-cab0-462f-9375-6266db0c5230 req-baad82cf-2a90-4736-ae3c-df06e3976d11 service nova] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Refreshing network info cache for port 6ce9925d-cf59-4de0-aece-da90fb4b9f36 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 859.979379] env[63088]: DEBUG oslo_vmware.api [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Task: {'id': task-1285029, 'name': PowerOnVM_Task} progress is 91%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.021902] env[63088]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-687a2a97-8aba-4c0d-86d4-2abd869d498e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.036914] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquiring lock "refresh_cache-9d5f1cde-e787-4a2c-ac49-83ec135ff51c" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.037168] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquired lock "refresh_cache-9d5f1cde-e787-4a2c-ac49-83ec135ff51c" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.037473] env[63088]: DEBUG nova.network.neutron [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 9d5f1cde-e787-4a2c-ac49-83ec135ff51c] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 860.071582] env[63088]: DEBUG oslo_vmware.api [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': task-1285030, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.518881} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.072296] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore2] 58d561fe-26aa-4e94-8d55-cc70c361b479/58d561fe-26aa-4e94-8d55-cc70c361b479.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 860.072792] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: 58d561fe-26aa-4e94-8d55-cc70c361b479] Extending root virtual disk to 1048576 {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 860.073601] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3e2bde7f-c7f2-4da4-8687-c4cd1d443727 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.081542] env[63088]: DEBUG oslo_vmware.api [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Waiting for the task: (returnval){ [ 860.081542] env[63088]: value = "task-1285032" [ 860.081542] env[63088]: _type = "Task" [ 860.081542] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.090785] env[63088]: DEBUG oslo_vmware.api [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': task-1285032, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.162136] env[63088]: DEBUG oslo_concurrency.lockutils [req-e6cca2d3-fae7-4ae8-8bb3-7eea7611e9da req-49365ffa-faa9-48b8-9cf5-a5898c986254 service nova] Releasing lock "refresh_cache-7b6aadb7-e34b-42b7-b69f-370434f5b665" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.410537] env[63088]: DEBUG oslo_vmware.api [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Task: {'id': task-1285029, 'name': PowerOnVM_Task, 'duration_secs': 1.04704} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.411144] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 860.411144] env[63088]: INFO nova.compute.manager [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Took 9.60 seconds to spawn the instance on the hypervisor. [ 860.411842] env[63088]: DEBUG nova.compute.manager [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 860.412349] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94311c48-ca11-4d26-8508-028c0ac93b86 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.581708] env[63088]: DEBUG nova.network.neutron [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 9d5f1cde-e787-4a2c-ac49-83ec135ff51c] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 860.599885] env[63088]: DEBUG oslo_vmware.api [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': task-1285032, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.102538} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.600357] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: 58d561fe-26aa-4e94-8d55-cc70c361b479] Extended root virtual disk {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 860.601324] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10378c52-e8bf-45a4-b9fb-36828351cefd {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.634099] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: 58d561fe-26aa-4e94-8d55-cc70c361b479] Reconfiguring VM instance instance-00000045 to attach disk [datastore2] 58d561fe-26aa-4e94-8d55-cc70c361b479/58d561fe-26aa-4e94-8d55-cc70c361b479.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 860.636862] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f59bef92-f014-4414-ae10-4ce3c31e9f68 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.662454] env[63088]: DEBUG oslo_vmware.api [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Waiting for the task: (returnval){ [ 860.662454] env[63088]: value = "task-1285033" [ 860.662454] env[63088]: _type = "Task" [ 860.662454] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.676774] env[63088]: DEBUG oslo_vmware.api [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': task-1285033, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.793521] env[63088]: DEBUG nova.network.neutron [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 9d5f1cde-e787-4a2c-ac49-83ec135ff51c] Updating instance_info_cache with network_info: [{"id": "07e1b344-3f50-473d-a443-caca3ba6c0ea", "address": "fa:16:3e:77:36:64", "network": {"id": "5b801381-6bb3-45cd-b81f-92b5d757e080", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1481457494-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e48f62dee9ad4e9b94b67bd871db5c63", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "051f343d-ac4f-4070-a26d-467603122c81", "external-id": "nsx-vlan-transportzone-277", "segmentation_id": 277, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap07e1b344-3f", "ovs_interfaceid": "07e1b344-3f50-473d-a443-caca3ba6c0ea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 860.849555] env[63088]: DEBUG nova.network.neutron [req-1e08c496-cab0-462f-9375-6266db0c5230 req-baad82cf-2a90-4736-ae3c-df06e3976d11 service nova] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Updated VIF entry in instance network info cache for port 6ce9925d-cf59-4de0-aece-da90fb4b9f36. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 860.850043] env[63088]: DEBUG nova.network.neutron [req-1e08c496-cab0-462f-9375-6266db0c5230 req-baad82cf-2a90-4736-ae3c-df06e3976d11 service nova] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Updating instance_info_cache with network_info: [{"id": "6ce9925d-cf59-4de0-aece-da90fb4b9f36", "address": "fa:16:3e:b0:ce:19", "network": {"id": "93b4f1f6-4529-4871-87dd-fbbe1fe59ea5", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-559913751-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.214", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7ed1e04b27ee464eb9a72866d8327352", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60badc2d-69d2-467d-a92e-98511f5cb0b2", "external-id": "cl2-zone-408", "segmentation_id": 408, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6ce9925d-cf", "ovs_interfaceid": "6ce9925d-cf59-4de0-aece-da90fb4b9f36", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 860.895816] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76e1894c-197e-4873-85b4-db5231b36bf4 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.904918] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63cdedbf-1b92-42c9-8727-3434e3609e0e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.950480] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44b8fae6-5042-4313-89f7-e7409f865439 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.952480] env[63088]: INFO nova.compute.manager [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Took 28.82 seconds to build instance. [ 860.959294] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf719c6c-8888-4738-8a51-29c09f81abe5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.975251] env[63088]: DEBUG nova.compute.provider_tree [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 861.173760] env[63088]: DEBUG oslo_vmware.api [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': task-1285033, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.297506] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Releasing lock "refresh_cache-9d5f1cde-e787-4a2c-ac49-83ec135ff51c" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 861.297871] env[63088]: DEBUG nova.compute.manager [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 9d5f1cde-e787-4a2c-ac49-83ec135ff51c] Instance network_info: |[{"id": "07e1b344-3f50-473d-a443-caca3ba6c0ea", "address": "fa:16:3e:77:36:64", "network": {"id": "5b801381-6bb3-45cd-b81f-92b5d757e080", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1481457494-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e48f62dee9ad4e9b94b67bd871db5c63", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "051f343d-ac4f-4070-a26d-467603122c81", "external-id": "nsx-vlan-transportzone-277", "segmentation_id": 277, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap07e1b344-3f", "ovs_interfaceid": "07e1b344-3f50-473d-a443-caca3ba6c0ea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 861.298745] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 9d5f1cde-e787-4a2c-ac49-83ec135ff51c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:77:36:64', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '051f343d-ac4f-4070-a26d-467603122c81', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '07e1b344-3f50-473d-a443-caca3ba6c0ea', 'vif_model': 'vmxnet3'}] {{(pid=63088) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 861.307716] env[63088]: DEBUG oslo.service.loopingcall [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 861.308778] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9d5f1cde-e787-4a2c-ac49-83ec135ff51c] Creating VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 861.309198] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6e2364d3-4a73-414d-a85e-c3774836ad6a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.331424] env[63088]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 861.331424] env[63088]: value = "task-1285034" [ 861.331424] env[63088]: _type = "Task" [ 861.331424] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.339960] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285034, 'name': CreateVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.352962] env[63088]: DEBUG oslo_concurrency.lockutils [req-1e08c496-cab0-462f-9375-6266db0c5230 req-baad82cf-2a90-4736-ae3c-df06e3976d11 service nova] Releasing lock "refresh_cache-dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 861.352962] env[63088]: DEBUG nova.compute.manager [req-1e08c496-cab0-462f-9375-6266db0c5230 req-baad82cf-2a90-4736-ae3c-df06e3976d11 service nova] [instance: 9d5f1cde-e787-4a2c-ac49-83ec135ff51c] Received event network-changed-07e1b344-3f50-473d-a443-caca3ba6c0ea {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 861.353306] env[63088]: DEBUG nova.compute.manager [req-1e08c496-cab0-462f-9375-6266db0c5230 req-baad82cf-2a90-4736-ae3c-df06e3976d11 service nova] [instance: 9d5f1cde-e787-4a2c-ac49-83ec135ff51c] Refreshing instance network info cache due to event network-changed-07e1b344-3f50-473d-a443-caca3ba6c0ea. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 861.353677] env[63088]: DEBUG oslo_concurrency.lockutils [req-1e08c496-cab0-462f-9375-6266db0c5230 req-baad82cf-2a90-4736-ae3c-df06e3976d11 service nova] Acquiring lock "refresh_cache-9d5f1cde-e787-4a2c-ac49-83ec135ff51c" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 861.353936] env[63088]: DEBUG oslo_concurrency.lockutils [req-1e08c496-cab0-462f-9375-6266db0c5230 req-baad82cf-2a90-4736-ae3c-df06e3976d11 service nova] Acquired lock "refresh_cache-9d5f1cde-e787-4a2c-ac49-83ec135ff51c" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 861.354263] env[63088]: DEBUG nova.network.neutron [req-1e08c496-cab0-462f-9375-6266db0c5230 req-baad82cf-2a90-4736-ae3c-df06e3976d11 service nova] [instance: 9d5f1cde-e787-4a2c-ac49-83ec135ff51c] Refreshing network info cache for port 07e1b344-3f50-473d-a443-caca3ba6c0ea {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 861.455517] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f90b8056-de0f-4a13-a8a8-52bd34a02ce7 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Lock "e6b0ce08-d67f-458f-92f5-1e904d03a5ef" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 53.895s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 861.479144] env[63088]: DEBUG nova.scheduler.client.report [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 861.675286] env[63088]: DEBUG oslo_vmware.api [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': task-1285033, 'name': ReconfigVM_Task, 'duration_secs': 0.691028} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.675286] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: 58d561fe-26aa-4e94-8d55-cc70c361b479] Reconfigured VM instance instance-00000045 to attach disk [datastore2] 58d561fe-26aa-4e94-8d55-cc70c361b479/58d561fe-26aa-4e94-8d55-cc70c361b479.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 861.675794] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e3892d27-f074-4e4f-9de0-b7ae59c8af92 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.683866] env[63088]: DEBUG oslo_vmware.api [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Waiting for the task: (returnval){ [ 861.683866] env[63088]: value = "task-1285035" [ 861.683866] env[63088]: _type = "Task" [ 861.683866] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.696714] env[63088]: DEBUG oslo_vmware.api [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': task-1285035, 'name': Rename_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.787457] env[63088]: DEBUG oslo_concurrency.lockutils [None req-ab25717c-7e13-4125-82df-f4c89863e7c3 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquiring lock "interface-4971b24c-6710-4f50-9846-727dad264b1f-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 861.788472] env[63088]: DEBUG oslo_concurrency.lockutils [None req-ab25717c-7e13-4125-82df-f4c89863e7c3 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Lock "interface-4971b24c-6710-4f50-9846-727dad264b1f-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 861.788472] env[63088]: DEBUG nova.objects.instance [None req-ab25717c-7e13-4125-82df-f4c89863e7c3 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Lazy-loading 'flavor' on Instance uuid 4971b24c-6710-4f50-9846-727dad264b1f {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 861.842369] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285034, 'name': CreateVM_Task, 'duration_secs': 0.429413} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.842583] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9d5f1cde-e787-4a2c-ac49-83ec135ff51c] Created VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 861.843354] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 861.843565] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 861.843935] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 861.844245] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b124e0dc-50a6-469a-9409-e2bbf6994b77 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.849267] env[63088]: DEBUG oslo_vmware.api [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 861.849267] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52798ef9-2831-1b67-0f1c-403d18c39b63" [ 861.849267] env[63088]: _type = "Task" [ 861.849267] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.860609] env[63088]: DEBUG oslo_vmware.api [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52798ef9-2831-1b67-0f1c-403d18c39b63, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.985361] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.512s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 861.986057] env[63088]: DEBUG nova.compute.manager [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 861.991530] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b548d275-1b16-4d7d-aec5-f1421ffffdea tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.057s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 861.991783] env[63088]: DEBUG nova.objects.instance [None req-b548d275-1b16-4d7d-aec5-f1421ffffdea tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Lazy-loading 'resources' on Instance uuid 2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 862.133541] env[63088]: DEBUG nova.network.neutron [req-1e08c496-cab0-462f-9375-6266db0c5230 req-baad82cf-2a90-4736-ae3c-df06e3976d11 service nova] [instance: 9d5f1cde-e787-4a2c-ac49-83ec135ff51c] Updated VIF entry in instance network info cache for port 07e1b344-3f50-473d-a443-caca3ba6c0ea. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 862.133910] env[63088]: DEBUG nova.network.neutron [req-1e08c496-cab0-462f-9375-6266db0c5230 req-baad82cf-2a90-4736-ae3c-df06e3976d11 service nova] [instance: 9d5f1cde-e787-4a2c-ac49-83ec135ff51c] Updating instance_info_cache with network_info: [{"id": "07e1b344-3f50-473d-a443-caca3ba6c0ea", "address": "fa:16:3e:77:36:64", "network": {"id": "5b801381-6bb3-45cd-b81f-92b5d757e080", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1481457494-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e48f62dee9ad4e9b94b67bd871db5c63", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "051f343d-ac4f-4070-a26d-467603122c81", "external-id": "nsx-vlan-transportzone-277", "segmentation_id": 277, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap07e1b344-3f", "ovs_interfaceid": "07e1b344-3f50-473d-a443-caca3ba6c0ea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 862.194322] env[63088]: DEBUG oslo_vmware.api [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': task-1285035, 'name': Rename_Task, 'duration_secs': 0.215407} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.194322] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: 58d561fe-26aa-4e94-8d55-cc70c361b479] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 862.194707] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-881da435-22c4-4d71-97db-9e56b137d5c5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.201684] env[63088]: DEBUG oslo_vmware.api [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Waiting for the task: (returnval){ [ 862.201684] env[63088]: value = "task-1285036" [ 862.201684] env[63088]: _type = "Task" [ 862.201684] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.210783] env[63088]: DEBUG oslo_vmware.api [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': task-1285036, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.362635] env[63088]: DEBUG oslo_vmware.api [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52798ef9-2831-1b67-0f1c-403d18c39b63, 'name': SearchDatastore_Task, 'duration_secs': 0.011917} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.362635] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 862.362635] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 9d5f1cde-e787-4a2c-ac49-83ec135ff51c] Processing image 1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 862.362635] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 862.363046] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.363046] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 862.363046] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6c8914e8-e9c2-4c3d-afd1-88aba6c38858 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.367385] env[63088]: DEBUG nova.objects.instance [None req-ab25717c-7e13-4125-82df-f4c89863e7c3 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Lazy-loading 'pci_requests' on Instance uuid 4971b24c-6710-4f50-9846-727dad264b1f {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 862.370690] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 862.370918] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 862.371755] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6fd7ffc5-f5ff-4b23-ac8e-417f321a5bea {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.378116] env[63088]: DEBUG oslo_vmware.api [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 862.378116] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52214bd8-7f4a-7058-aee3-d15e66188583" [ 862.378116] env[63088]: _type = "Task" [ 862.378116] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.387995] env[63088]: DEBUG oslo_vmware.api [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52214bd8-7f4a-7058-aee3-d15e66188583, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.498520] env[63088]: DEBUG nova.compute.utils [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 862.500032] env[63088]: DEBUG nova.compute.manager [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 862.500204] env[63088]: DEBUG nova.network.neutron [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 862.542990] env[63088]: DEBUG nova.policy [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4424adc5bbe34a91b669c3a907882644', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f33f2701fad94864a8c406a404bc0a42', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 862.637447] env[63088]: DEBUG oslo_concurrency.lockutils [req-1e08c496-cab0-462f-9375-6266db0c5230 req-baad82cf-2a90-4736-ae3c-df06e3976d11 service nova] Releasing lock "refresh_cache-9d5f1cde-e787-4a2c-ac49-83ec135ff51c" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 862.718075] env[63088]: DEBUG oslo_vmware.api [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': task-1285036, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.817476] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e10c4623-797f-49e3-9e4b-30cb38520b16 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.826236] env[63088]: DEBUG nova.network.neutron [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Successfully created port: 1fb0e170-03ab-4ee4-9f1f-c5c6008c823d {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 862.828996] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cd1f845-c95b-4403-840c-5a89b7add133 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.872236] env[63088]: DEBUG nova.objects.base [None req-ab25717c-7e13-4125-82df-f4c89863e7c3 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Object Instance<4971b24c-6710-4f50-9846-727dad264b1f> lazy-loaded attributes: flavor,pci_requests {{(pid=63088) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 862.873430] env[63088]: DEBUG nova.network.neutron [None req-ab25717c-7e13-4125-82df-f4c89863e7c3 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 862.877224] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93f835b5-b5b8-4e52-b99f-6f43a41efb92 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.898807] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70c74db5-4073-468f-8167-3c7285347f85 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.903333] env[63088]: DEBUG oslo_vmware.api [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52214bd8-7f4a-7058-aee3-d15e66188583, 'name': SearchDatastore_Task, 'duration_secs': 0.012413} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.904614] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-70f946d4-421c-4dd3-99a8-9fc5f18d74ad {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.915163] env[63088]: DEBUG nova.compute.provider_tree [None req-b548d275-1b16-4d7d-aec5-f1421ffffdea tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 862.920367] env[63088]: DEBUG oslo_vmware.api [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 862.920367] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]5276d33b-d800-7e52-e0c5-87bd591de705" [ 862.920367] env[63088]: _type = "Task" [ 862.920367] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.928501] env[63088]: DEBUG oslo_vmware.api [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]5276d33b-d800-7e52-e0c5-87bd591de705, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.965789] env[63088]: DEBUG nova.policy [None req-ab25717c-7e13-4125-82df-f4c89863e7c3 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '93f1d334e71a4a99ba300c3ee6d70243', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'aa3d24a1a6c0430985fd80365d986ee1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 862.969845] env[63088]: DEBUG nova.compute.manager [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Stashing vm_state: active {{(pid=63088) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 863.004224] env[63088]: DEBUG nova.compute.manager [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 863.214411] env[63088]: DEBUG oslo_vmware.api [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': task-1285036, 'name': PowerOnVM_Task, 'duration_secs': 0.535108} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.214411] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: 58d561fe-26aa-4e94-8d55-cc70c361b479] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 863.214566] env[63088]: INFO nova.compute.manager [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: 58d561fe-26aa-4e94-8d55-cc70c361b479] Took 8.11 seconds to spawn the instance on the hypervisor. [ 863.214620] env[63088]: DEBUG nova.compute.manager [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: 58d561fe-26aa-4e94-8d55-cc70c361b479] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 863.215425] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d886ed3f-33dd-4c0b-b4e4-8c507b2ad9ee {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.320046] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 863.320303] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 863.355573] env[63088]: DEBUG nova.network.neutron [None req-ab25717c-7e13-4125-82df-f4c89863e7c3 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Successfully created port: 81932288-bc32-47c4-a3b2-02597cb6bf81 {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 863.419208] env[63088]: DEBUG nova.scheduler.client.report [None req-b548d275-1b16-4d7d-aec5-f1421ffffdea tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 863.434460] env[63088]: DEBUG oslo_vmware.api [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]5276d33b-d800-7e52-e0c5-87bd591de705, 'name': SearchDatastore_Task, 'duration_secs': 0.012163} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.434944] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 863.435271] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore2] 9d5f1cde-e787-4a2c-ac49-83ec135ff51c/9d5f1cde-e787-4a2c-ac49-83ec135ff51c.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 863.435505] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2aca79d6-f7f3-43ec-befd-e039788f10e4 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.443258] env[63088]: DEBUG oslo_vmware.api [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 863.443258] env[63088]: value = "task-1285037" [ 863.443258] env[63088]: _type = "Task" [ 863.443258] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.454431] env[63088]: DEBUG oslo_vmware.api [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285037, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.493363] env[63088]: DEBUG oslo_concurrency.lockutils [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.740653] env[63088]: INFO nova.compute.manager [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: 58d561fe-26aa-4e94-8d55-cc70c361b479] Took 30.49 seconds to build instance. [ 863.829208] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 863.829821] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Starting heal instance info cache {{(pid=63088) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 863.829821] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Rebuilding the list of instances to heal {{(pid=63088) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 863.929449] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b548d275-1b16-4d7d-aec5-f1421ffffdea tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.938s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.932953] env[63088]: DEBUG oslo_concurrency.lockutils [None req-8102b1ea-81e5-4196-b4f7-a4fd361924a7 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.195s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.933388] env[63088]: DEBUG nova.objects.instance [None req-8102b1ea-81e5-4196-b4f7-a4fd361924a7 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Lazy-loading 'resources' on Instance uuid b159649b-ebd5-4c7d-9074-dc0e6395a947 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 863.955326] env[63088]: DEBUG oslo_vmware.api [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285037, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.956410] env[63088]: INFO nova.scheduler.client.report [None req-b548d275-1b16-4d7d-aec5-f1421ffffdea tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Deleted allocations for instance 2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68 [ 864.014529] env[63088]: DEBUG nova.compute.manager [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 864.044444] env[63088]: DEBUG nova.virt.hardware [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 864.044597] env[63088]: DEBUG nova.virt.hardware [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 864.044848] env[63088]: DEBUG nova.virt.hardware [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 864.045075] env[63088]: DEBUG nova.virt.hardware [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 864.045256] env[63088]: DEBUG nova.virt.hardware [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 864.045422] env[63088]: DEBUG nova.virt.hardware [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 864.045656] env[63088]: DEBUG nova.virt.hardware [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 864.045848] env[63088]: DEBUG nova.virt.hardware [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 864.046076] env[63088]: DEBUG nova.virt.hardware [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 864.046273] env[63088]: DEBUG nova.virt.hardware [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 864.046469] env[63088]: DEBUG nova.virt.hardware [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 864.047419] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b3fc665-dae2-49e3-a7c3-952f5fa94132 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.057490] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6ed51da-7664-480d-83c1-04a574e0e06d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.243298] env[63088]: DEBUG oslo_concurrency.lockutils [None req-542ee47a-3371-4d4b-b14e-7d27a7cc0bea tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Lock "58d561fe-26aa-4e94-8d55-cc70c361b479" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 40.364s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.336295] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: 9d5f1cde-e787-4a2c-ac49-83ec135ff51c] Skipping network cache update for instance because it is Building. {{(pid=63088) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 864.336508] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Skipping network cache update for instance because it is Building. {{(pid=63088) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 864.381518] env[63088]: DEBUG nova.network.neutron [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Successfully updated port: 1fb0e170-03ab-4ee4-9f1f-c5c6008c823d {{(pid=63088) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 864.387866] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Acquiring lock "refresh_cache-db032da3-d334-486f-a428-af1c8a3c360a" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 864.388722] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Acquired lock "refresh_cache-db032da3-d334-486f-a428-af1c8a3c360a" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.388722] env[63088]: DEBUG nova.network.neutron [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: db032da3-d334-486f-a428-af1c8a3c360a] Forcefully refreshing network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 864.388722] env[63088]: DEBUG nova.objects.instance [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Lazy-loading 'info_cache' on Instance uuid db032da3-d334-486f-a428-af1c8a3c360a {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 864.455928] env[63088]: DEBUG oslo_vmware.api [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285037, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.823315} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.456219] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore2] 9d5f1cde-e787-4a2c-ac49-83ec135ff51c/9d5f1cde-e787-4a2c-ac49-83ec135ff51c.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 864.456463] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 9d5f1cde-e787-4a2c-ac49-83ec135ff51c] Extending root virtual disk to 1048576 {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 864.456729] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c3c1ead6-1af4-439e-8991-43a32e96459c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.469514] env[63088]: DEBUG oslo_vmware.api [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 864.469514] env[63088]: value = "task-1285038" [ 864.469514] env[63088]: _type = "Task" [ 864.469514] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.470018] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b548d275-1b16-4d7d-aec5-f1421ffffdea tempest-ServerPasswordTestJSON-1303384023 tempest-ServerPasswordTestJSON-1303384023-project-member] Lock "2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.519s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.480035] env[63088]: DEBUG oslo_vmware.api [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285038, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.527117] env[63088]: DEBUG nova.compute.manager [req-4f0f6263-57d2-460f-94df-706cdc8b65ec req-5f647c2e-9795-416b-ac71-8782455fa576 service nova] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Received event network-vif-plugged-1fb0e170-03ab-4ee4-9f1f-c5c6008c823d {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 864.527455] env[63088]: DEBUG oslo_concurrency.lockutils [req-4f0f6263-57d2-460f-94df-706cdc8b65ec req-5f647c2e-9795-416b-ac71-8782455fa576 service nova] Acquiring lock "2269e520-08b6-433f-8fe9-7b84d91e02d7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 864.527704] env[63088]: DEBUG oslo_concurrency.lockutils [req-4f0f6263-57d2-460f-94df-706cdc8b65ec req-5f647c2e-9795-416b-ac71-8782455fa576 service nova] Lock "2269e520-08b6-433f-8fe9-7b84d91e02d7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 864.527935] env[63088]: DEBUG oslo_concurrency.lockutils [req-4f0f6263-57d2-460f-94df-706cdc8b65ec req-5f647c2e-9795-416b-ac71-8782455fa576 service nova] Lock "2269e520-08b6-433f-8fe9-7b84d91e02d7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.528165] env[63088]: DEBUG nova.compute.manager [req-4f0f6263-57d2-460f-94df-706cdc8b65ec req-5f647c2e-9795-416b-ac71-8782455fa576 service nova] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] No waiting events found dispatching network-vif-plugged-1fb0e170-03ab-4ee4-9f1f-c5c6008c823d {{(pid=63088) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 864.528376] env[63088]: WARNING nova.compute.manager [req-4f0f6263-57d2-460f-94df-706cdc8b65ec req-5f647c2e-9795-416b-ac71-8782455fa576 service nova] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Received unexpected event network-vif-plugged-1fb0e170-03ab-4ee4-9f1f-c5c6008c823d for instance with vm_state building and task_state spawning. [ 864.692246] env[63088]: INFO nova.compute.manager [None req-058a9ba6-53de-4c1c-9276-bca158f57395 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: 58d561fe-26aa-4e94-8d55-cc70c361b479] Rescuing [ 864.692510] env[63088]: DEBUG oslo_concurrency.lockutils [None req-058a9ba6-53de-4c1c-9276-bca158f57395 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Acquiring lock "refresh_cache-58d561fe-26aa-4e94-8d55-cc70c361b479" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 864.692666] env[63088]: DEBUG oslo_concurrency.lockutils [None req-058a9ba6-53de-4c1c-9276-bca158f57395 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Acquired lock "refresh_cache-58d561fe-26aa-4e94-8d55-cc70c361b479" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.692894] env[63088]: DEBUG nova.network.neutron [None req-058a9ba6-53de-4c1c-9276-bca158f57395 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: 58d561fe-26aa-4e94-8d55-cc70c361b479] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 864.734985] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26e27c04-00c7-423a-8900-6e102107244f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.744182] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-701d00d6-4aa1-40e1-8a77-aa6ac568cd24 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.779816] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99662f5f-8f57-4a04-aa98-cffd19499220 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.788420] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a519591e-ea48-4a0e-89a6-ff18c4a1506f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.805182] env[63088]: DEBUG nova.compute.provider_tree [None req-8102b1ea-81e5-4196-b4f7-a4fd361924a7 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 864.884314] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquiring lock "refresh_cache-2269e520-08b6-433f-8fe9-7b84d91e02d7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 864.884460] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquired lock "refresh_cache-2269e520-08b6-433f-8fe9-7b84d91e02d7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.885283] env[63088]: DEBUG nova.network.neutron [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 864.981826] env[63088]: DEBUG oslo_vmware.api [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285038, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.103253} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.982248] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 9d5f1cde-e787-4a2c-ac49-83ec135ff51c] Extended root virtual disk {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 864.983497] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52f903f4-e02e-4788-9d56-7e0553ea35d7 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.013856] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 9d5f1cde-e787-4a2c-ac49-83ec135ff51c] Reconfiguring VM instance instance-00000046 to attach disk [datastore2] 9d5f1cde-e787-4a2c-ac49-83ec135ff51c/9d5f1cde-e787-4a2c-ac49-83ec135ff51c.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 865.013856] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dd65208a-e26c-4a2d-9150-9e0867e8c571 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.037757] env[63088]: DEBUG oslo_vmware.api [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 865.037757] env[63088]: value = "task-1285039" [ 865.037757] env[63088]: _type = "Task" [ 865.037757] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.046228] env[63088]: DEBUG oslo_vmware.api [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285039, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.120335] env[63088]: DEBUG nova.network.neutron [None req-ab25717c-7e13-4125-82df-f4c89863e7c3 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Successfully updated port: 81932288-bc32-47c4-a3b2-02597cb6bf81 {{(pid=63088) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 865.141656] env[63088]: DEBUG nova.compute.manager [req-275dd459-9651-42a4-acc1-6c3ab35e1a6c req-92c507ce-ffce-4392-913c-7f0e3857e11f service nova] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Received event network-vif-plugged-81932288-bc32-47c4-a3b2-02597cb6bf81 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 865.141950] env[63088]: DEBUG oslo_concurrency.lockutils [req-275dd459-9651-42a4-acc1-6c3ab35e1a6c req-92c507ce-ffce-4392-913c-7f0e3857e11f service nova] Acquiring lock "4971b24c-6710-4f50-9846-727dad264b1f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 865.142297] env[63088]: DEBUG oslo_concurrency.lockutils [req-275dd459-9651-42a4-acc1-6c3ab35e1a6c req-92c507ce-ffce-4392-913c-7f0e3857e11f service nova] Lock "4971b24c-6710-4f50-9846-727dad264b1f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 865.142510] env[63088]: DEBUG oslo_concurrency.lockutils [req-275dd459-9651-42a4-acc1-6c3ab35e1a6c req-92c507ce-ffce-4392-913c-7f0e3857e11f service nova] Lock "4971b24c-6710-4f50-9846-727dad264b1f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.142731] env[63088]: DEBUG nova.compute.manager [req-275dd459-9651-42a4-acc1-6c3ab35e1a6c req-92c507ce-ffce-4392-913c-7f0e3857e11f service nova] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] No waiting events found dispatching network-vif-plugged-81932288-bc32-47c4-a3b2-02597cb6bf81 {{(pid=63088) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 865.147014] env[63088]: WARNING nova.compute.manager [req-275dd459-9651-42a4-acc1-6c3ab35e1a6c req-92c507ce-ffce-4392-913c-7f0e3857e11f service nova] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Received unexpected event network-vif-plugged-81932288-bc32-47c4-a3b2-02597cb6bf81 for instance with vm_state active and task_state None. [ 865.308902] env[63088]: DEBUG nova.scheduler.client.report [None req-8102b1ea-81e5-4196-b4f7-a4fd361924a7 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 865.450511] env[63088]: DEBUG nova.network.neutron [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 865.503756] env[63088]: DEBUG nova.network.neutron [None req-058a9ba6-53de-4c1c-9276-bca158f57395 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: 58d561fe-26aa-4e94-8d55-cc70c361b479] Updating instance_info_cache with network_info: [{"id": "c6477c64-3129-4546-84b1-40bc79bcaa57", "address": "fa:16:3e:75:e7:1a", "network": {"id": "7db9302e-1c7a-42dc-8084-98e2a57ad7bf", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-2073161169-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "900466c3c09a4f928bdd4b602715c838", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "87bbf4e0-9064-4516-b7e7-44973f817205", "external-id": "nsx-vlan-transportzone-507", "segmentation_id": 507, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6477c64-31", "ovs_interfaceid": "c6477c64-3129-4546-84b1-40bc79bcaa57", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 865.561471] env[63088]: DEBUG oslo_vmware.api [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285039, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.625046] env[63088]: DEBUG oslo_concurrency.lockutils [None req-ab25717c-7e13-4125-82df-f4c89863e7c3 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquiring lock "refresh_cache-4971b24c-6710-4f50-9846-727dad264b1f" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 865.625265] env[63088]: DEBUG oslo_concurrency.lockutils [None req-ab25717c-7e13-4125-82df-f4c89863e7c3 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquired lock "refresh_cache-4971b24c-6710-4f50-9846-727dad264b1f" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 865.625453] env[63088]: DEBUG nova.network.neutron [None req-ab25717c-7e13-4125-82df-f4c89863e7c3 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 865.676786] env[63088]: DEBUG nova.network.neutron [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Updating instance_info_cache with network_info: [{"id": "1fb0e170-03ab-4ee4-9f1f-c5c6008c823d", "address": "fa:16:3e:a8:02:36", "network": {"id": "dff14a7f-0af1-4e4e-a498-86d7c9816e6b", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-762073671-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f33f2701fad94864a8c406a404bc0a42", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "21310d90-efbc-45a8-a97f-c4358606530f", "external-id": "nsx-vlan-transportzone-672", "segmentation_id": 672, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1fb0e170-03", "ovs_interfaceid": "1fb0e170-03ab-4ee4-9f1f-c5c6008c823d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 865.814290] env[63088]: DEBUG oslo_concurrency.lockutils [None req-8102b1ea-81e5-4196-b4f7-a4fd361924a7 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.881s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.816807] env[63088]: DEBUG oslo_concurrency.lockutils [None req-ad047aea-ed6a-4a24-b3ca-46aa3078b110 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.477s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 865.817061] env[63088]: DEBUG nova.objects.instance [None req-ad047aea-ed6a-4a24-b3ca-46aa3078b110 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Lazy-loading 'resources' on Instance uuid 07fd3e6f-e0af-467c-9039-238be5c58d25 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 865.838085] env[63088]: INFO nova.scheduler.client.report [None req-8102b1ea-81e5-4196-b4f7-a4fd361924a7 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Deleted allocations for instance b159649b-ebd5-4c7d-9074-dc0e6395a947 [ 866.009837] env[63088]: DEBUG oslo_concurrency.lockutils [None req-058a9ba6-53de-4c1c-9276-bca158f57395 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Releasing lock "refresh_cache-58d561fe-26aa-4e94-8d55-cc70c361b479" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 866.056700] env[63088]: DEBUG oslo_vmware.api [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285039, 'name': ReconfigVM_Task, 'duration_secs': 0.566711} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.057655] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 9d5f1cde-e787-4a2c-ac49-83ec135ff51c] Reconfigured VM instance instance-00000046 to attach disk [datastore2] 9d5f1cde-e787-4a2c-ac49-83ec135ff51c/9d5f1cde-e787-4a2c-ac49-83ec135ff51c.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 866.058389] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bdedb460-7847-4073-b7d2-ef0b4669d7c3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.065704] env[63088]: DEBUG oslo_vmware.api [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 866.065704] env[63088]: value = "task-1285040" [ 866.065704] env[63088]: _type = "Task" [ 866.065704] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.074693] env[63088]: DEBUG oslo_vmware.api [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285040, 'name': Rename_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.183251] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Releasing lock "refresh_cache-2269e520-08b6-433f-8fe9-7b84d91e02d7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 866.186396] env[63088]: DEBUG nova.compute.manager [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Instance network_info: |[{"id": "1fb0e170-03ab-4ee4-9f1f-c5c6008c823d", "address": "fa:16:3e:a8:02:36", "network": {"id": "dff14a7f-0af1-4e4e-a498-86d7c9816e6b", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-762073671-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f33f2701fad94864a8c406a404bc0a42", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "21310d90-efbc-45a8-a97f-c4358606530f", "external-id": "nsx-vlan-transportzone-672", "segmentation_id": 672, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1fb0e170-03", "ovs_interfaceid": "1fb0e170-03ab-4ee4-9f1f-c5c6008c823d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 866.186574] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a8:02:36', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '21310d90-efbc-45a8-a97f-c4358606530f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1fb0e170-03ab-4ee4-9f1f-c5c6008c823d', 'vif_model': 'vmxnet3'}] {{(pid=63088) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 866.196025] env[63088]: DEBUG oslo.service.loopingcall [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 866.196898] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Creating VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 866.197188] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e6dc44f3-a847-4e42-816a-bfff438d95dd {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.218726] env[63088]: WARNING nova.network.neutron [None req-ab25717c-7e13-4125-82df-f4c89863e7c3 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] 1174a284-d3ca-4f9e-aa81-13ee9a693e55 already exists in list: networks containing: ['1174a284-d3ca-4f9e-aa81-13ee9a693e55']. ignoring it [ 866.226166] env[63088]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 866.226166] env[63088]: value = "task-1285041" [ 866.226166] env[63088]: _type = "Task" [ 866.226166] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.236715] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285041, 'name': CreateVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.341515] env[63088]: DEBUG nova.scheduler.client.report [None req-ad047aea-ed6a-4a24-b3ca-46aa3078b110 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Refreshing inventories for resource provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 866.347421] env[63088]: DEBUG oslo_concurrency.lockutils [None req-8102b1ea-81e5-4196-b4f7-a4fd361924a7 tempest-ServerShowV257Test-1730712990 tempest-ServerShowV257Test-1730712990-project-member] Lock "b159649b-ebd5-4c7d-9074-dc0e6395a947" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.335s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.375408] env[63088]: DEBUG nova.scheduler.client.report [None req-ad047aea-ed6a-4a24-b3ca-46aa3078b110 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Updating ProviderTree inventory for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 866.375644] env[63088]: DEBUG nova.compute.provider_tree [None req-ad047aea-ed6a-4a24-b3ca-46aa3078b110 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Updating inventory in ProviderTree for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 866.380551] env[63088]: DEBUG nova.network.neutron [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: db032da3-d334-486f-a428-af1c8a3c360a] Updating instance_info_cache with network_info: [{"id": "98219984-3198-4e54-8c93-024446f958a3", "address": "fa:16:3e:6d:e6:fe", "network": {"id": "7db9302e-1c7a-42dc-8084-98e2a57ad7bf", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-2073161169-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "900466c3c09a4f928bdd4b602715c838", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "87bbf4e0-9064-4516-b7e7-44973f817205", "external-id": "nsx-vlan-transportzone-507", "segmentation_id": 507, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap98219984-31", "ovs_interfaceid": "98219984-3198-4e54-8c93-024446f958a3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 866.399973] env[63088]: DEBUG nova.scheduler.client.report [None req-ad047aea-ed6a-4a24-b3ca-46aa3078b110 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Refreshing aggregate associations for resource provider 6eae54a9-8831-40eb-bf54-4bc60d346b02, aggregates: None {{(pid=63088) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 866.422559] env[63088]: DEBUG nova.scheduler.client.report [None req-ad047aea-ed6a-4a24-b3ca-46aa3078b110 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Refreshing trait associations for resource provider 6eae54a9-8831-40eb-bf54-4bc60d346b02, traits: COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO {{(pid=63088) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 866.556674] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-058a9ba6-53de-4c1c-9276-bca158f57395 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: 58d561fe-26aa-4e94-8d55-cc70c361b479] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 866.556963] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1164e5dc-fa97-4f73-9764-368618d42e89 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.566876] env[63088]: DEBUG oslo_vmware.api [None req-058a9ba6-53de-4c1c-9276-bca158f57395 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Waiting for the task: (returnval){ [ 866.566876] env[63088]: value = "task-1285042" [ 866.566876] env[63088]: _type = "Task" [ 866.566876] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.582866] env[63088]: DEBUG oslo_vmware.api [None req-058a9ba6-53de-4c1c-9276-bca158f57395 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': task-1285042, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.588850] env[63088]: DEBUG oslo_vmware.api [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285040, 'name': Rename_Task, 'duration_secs': 0.267322} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.591416] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 9d5f1cde-e787-4a2c-ac49-83ec135ff51c] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 866.591753] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c69a5b22-cc64-4858-af07-2462bb538b35 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.597815] env[63088]: DEBUG oslo_vmware.api [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 866.597815] env[63088]: value = "task-1285043" [ 866.597815] env[63088]: _type = "Task" [ 866.597815] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.606369] env[63088]: DEBUG oslo_vmware.api [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285043, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.737216] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285041, 'name': CreateVM_Task, 'duration_secs': 0.355643} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.742377] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Created VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 866.742377] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 866.742377] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 866.742377] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 866.742377] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-439360ae-6e86-48d7-b18d-4a28500daabf {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.749280] env[63088]: DEBUG oslo_vmware.api [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Waiting for the task: (returnval){ [ 866.749280] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]525303fe-2f57-90ac-4e6e-506a89c8eef7" [ 866.749280] env[63088]: _type = "Task" [ 866.749280] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.755356] env[63088]: DEBUG nova.compute.manager [req-90a8a8cb-f252-43cd-847e-645d52a306ee req-d014781a-ca21-4b51-97ff-f122b03e498e service nova] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Received event network-changed-1fb0e170-03ab-4ee4-9f1f-c5c6008c823d {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 866.755356] env[63088]: DEBUG nova.compute.manager [req-90a8a8cb-f252-43cd-847e-645d52a306ee req-d014781a-ca21-4b51-97ff-f122b03e498e service nova] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Refreshing instance network info cache due to event network-changed-1fb0e170-03ab-4ee4-9f1f-c5c6008c823d. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 866.755356] env[63088]: DEBUG oslo_concurrency.lockutils [req-90a8a8cb-f252-43cd-847e-645d52a306ee req-d014781a-ca21-4b51-97ff-f122b03e498e service nova] Acquiring lock "refresh_cache-2269e520-08b6-433f-8fe9-7b84d91e02d7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 866.755576] env[63088]: DEBUG oslo_concurrency.lockutils [req-90a8a8cb-f252-43cd-847e-645d52a306ee req-d014781a-ca21-4b51-97ff-f122b03e498e service nova] Acquired lock "refresh_cache-2269e520-08b6-433f-8fe9-7b84d91e02d7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 866.755702] env[63088]: DEBUG nova.network.neutron [req-90a8a8cb-f252-43cd-847e-645d52a306ee req-d014781a-ca21-4b51-97ff-f122b03e498e service nova] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Refreshing network info cache for port 1fb0e170-03ab-4ee4-9f1f-c5c6008c823d {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 866.765294] env[63088]: DEBUG oslo_vmware.api [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]525303fe-2f57-90ac-4e6e-506a89c8eef7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.793166] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04e45dad-4459-4d95-b16d-ea06fce334e9 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.801475] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a1af2ba-9a6e-4ee8-809c-210cf2b642e6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.807137] env[63088]: DEBUG nova.network.neutron [None req-ab25717c-7e13-4125-82df-f4c89863e7c3 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Updating instance_info_cache with network_info: [{"id": "d5a10951-ae2e-48f5-b0ca-b1f144e5fe3a", "address": "fa:16:3e:95:96:0c", "network": {"id": "1174a284-d3ca-4f9e-aa81-13ee9a693e55", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1994276040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.166", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa3d24a1a6c0430985fd80365d986ee1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd5a10951-ae", "ovs_interfaceid": "d5a10951-ae2e-48f5-b0ca-b1f144e5fe3a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "81932288-bc32-47c4-a3b2-02597cb6bf81", "address": "fa:16:3e:17:d1:1a", "network": {"id": "1174a284-d3ca-4f9e-aa81-13ee9a693e55", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1994276040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa3d24a1a6c0430985fd80365d986ee1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap81932288-bc", "ovs_interfaceid": "81932288-bc32-47c4-a3b2-02597cb6bf81", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 866.836876] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee2bb2ef-6898-4516-b18c-5010471458ac {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.845474] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7455dfc8-6bc4-4927-b684-eb5d3b7767cf {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.860149] env[63088]: DEBUG nova.compute.provider_tree [None req-ad047aea-ed6a-4a24-b3ca-46aa3078b110 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 866.890355] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Releasing lock "refresh_cache-db032da3-d334-486f-a428-af1c8a3c360a" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 866.890600] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: db032da3-d334-486f-a428-af1c8a3c360a] Updated the network info_cache for instance {{(pid=63088) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 866.891227] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 866.891411] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 866.891617] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 866.891719] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 866.891854] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 866.891991] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 866.892134] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63088) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 866.892274] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager.update_available_resource {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 867.082195] env[63088]: DEBUG oslo_vmware.api [None req-058a9ba6-53de-4c1c-9276-bca158f57395 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': task-1285042, 'name': PowerOffVM_Task, 'duration_secs': 0.305583} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.082505] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-058a9ba6-53de-4c1c-9276-bca158f57395 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: 58d561fe-26aa-4e94-8d55-cc70c361b479] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 867.083317] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7c9bded-fc9c-448c-a6a9-b7f03c247504 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.104014] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47793caa-3783-42b8-9d8c-c1380ff03297 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.116500] env[63088]: DEBUG oslo_vmware.api [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285043, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.138800] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-058a9ba6-53de-4c1c-9276-bca158f57395 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: 58d561fe-26aa-4e94-8d55-cc70c361b479] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 867.139148] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d0a2fa3b-2da9-423e-816d-2892be246db6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.146192] env[63088]: DEBUG oslo_vmware.api [None req-058a9ba6-53de-4c1c-9276-bca158f57395 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Waiting for the task: (returnval){ [ 867.146192] env[63088]: value = "task-1285044" [ 867.146192] env[63088]: _type = "Task" [ 867.146192] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.156619] env[63088]: DEBUG oslo_vmware.api [None req-058a9ba6-53de-4c1c-9276-bca158f57395 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': task-1285044, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.257802] env[63088]: DEBUG oslo_vmware.api [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]525303fe-2f57-90ac-4e6e-506a89c8eef7, 'name': SearchDatastore_Task, 'duration_secs': 0.01752} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.258089] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 867.258327] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Processing image 1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 867.258559] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 867.258705] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.258882] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 867.259169] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-952845a5-f898-43f3-ad51-35918f6ecb10 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.270095] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 867.270336] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 867.271440] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3a72218e-08d5-4da0-8d4d-19e134c886d7 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.277283] env[63088]: DEBUG oslo_vmware.api [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Waiting for the task: (returnval){ [ 867.277283] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52ef9333-bd23-1d45-1e8b-037185bbdc23" [ 867.277283] env[63088]: _type = "Task" [ 867.277283] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.286668] env[63088]: DEBUG oslo_vmware.api [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52ef9333-bd23-1d45-1e8b-037185bbdc23, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.295105] env[63088]: DEBUG nova.compute.manager [req-4a66b9bc-2308-4094-909f-c594cad43562 req-dade105e-15a5-4445-9b6f-e2746d22839b service nova] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Received event network-changed-81932288-bc32-47c4-a3b2-02597cb6bf81 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 867.295313] env[63088]: DEBUG nova.compute.manager [req-4a66b9bc-2308-4094-909f-c594cad43562 req-dade105e-15a5-4445-9b6f-e2746d22839b service nova] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Refreshing instance network info cache due to event network-changed-81932288-bc32-47c4-a3b2-02597cb6bf81. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 867.295515] env[63088]: DEBUG oslo_concurrency.lockutils [req-4a66b9bc-2308-4094-909f-c594cad43562 req-dade105e-15a5-4445-9b6f-e2746d22839b service nova] Acquiring lock "refresh_cache-4971b24c-6710-4f50-9846-727dad264b1f" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 867.310304] env[63088]: DEBUG oslo_concurrency.lockutils [None req-ab25717c-7e13-4125-82df-f4c89863e7c3 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Releasing lock "refresh_cache-4971b24c-6710-4f50-9846-727dad264b1f" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 867.311030] env[63088]: DEBUG oslo_concurrency.lockutils [None req-ab25717c-7e13-4125-82df-f4c89863e7c3 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquiring lock "4971b24c-6710-4f50-9846-727dad264b1f" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 867.311311] env[63088]: DEBUG oslo_concurrency.lockutils [None req-ab25717c-7e13-4125-82df-f4c89863e7c3 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquired lock "4971b24c-6710-4f50-9846-727dad264b1f" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.311598] env[63088]: DEBUG oslo_concurrency.lockutils [req-4a66b9bc-2308-4094-909f-c594cad43562 req-dade105e-15a5-4445-9b6f-e2746d22839b service nova] Acquired lock "refresh_cache-4971b24c-6710-4f50-9846-727dad264b1f" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.311788] env[63088]: DEBUG nova.network.neutron [req-4a66b9bc-2308-4094-909f-c594cad43562 req-dade105e-15a5-4445-9b6f-e2746d22839b service nova] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Refreshing network info cache for port 81932288-bc32-47c4-a3b2-02597cb6bf81 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 867.314260] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ce06aea-c328-48e4-ae66-784f1b6efc48 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.333174] env[63088]: DEBUG nova.virt.hardware [None req-ab25717c-7e13-4125-82df-f4c89863e7c3 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 867.334033] env[63088]: DEBUG nova.virt.hardware [None req-ab25717c-7e13-4125-82df-f4c89863e7c3 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 867.334033] env[63088]: DEBUG nova.virt.hardware [None req-ab25717c-7e13-4125-82df-f4c89863e7c3 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 867.334033] env[63088]: DEBUG nova.virt.hardware [None req-ab25717c-7e13-4125-82df-f4c89863e7c3 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 867.334033] env[63088]: DEBUG nova.virt.hardware [None req-ab25717c-7e13-4125-82df-f4c89863e7c3 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 867.334252] env[63088]: DEBUG nova.virt.hardware [None req-ab25717c-7e13-4125-82df-f4c89863e7c3 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 867.334305] env[63088]: DEBUG nova.virt.hardware [None req-ab25717c-7e13-4125-82df-f4c89863e7c3 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 867.334598] env[63088]: DEBUG nova.virt.hardware [None req-ab25717c-7e13-4125-82df-f4c89863e7c3 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 867.334675] env[63088]: DEBUG nova.virt.hardware [None req-ab25717c-7e13-4125-82df-f4c89863e7c3 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 867.334778] env[63088]: DEBUG nova.virt.hardware [None req-ab25717c-7e13-4125-82df-f4c89863e7c3 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 867.334949] env[63088]: DEBUG nova.virt.hardware [None req-ab25717c-7e13-4125-82df-f4c89863e7c3 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 867.341971] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-ab25717c-7e13-4125-82df-f4c89863e7c3 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Reconfiguring VM to attach interface {{(pid=63088) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 867.342695] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-59e037a4-ee7d-4dad-b4e6-cef98d86357e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.364495] env[63088]: DEBUG nova.scheduler.client.report [None req-ad047aea-ed6a-4a24-b3ca-46aa3078b110 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 867.367794] env[63088]: DEBUG oslo_vmware.api [None req-ab25717c-7e13-4125-82df-f4c89863e7c3 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Waiting for the task: (returnval){ [ 867.367794] env[63088]: value = "task-1285045" [ 867.367794] env[63088]: _type = "Task" [ 867.367794] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.377301] env[63088]: DEBUG oslo_vmware.api [None req-ab25717c-7e13-4125-82df-f4c89863e7c3 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285045, 'name': ReconfigVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.395221] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 867.548973] env[63088]: DEBUG nova.network.neutron [req-90a8a8cb-f252-43cd-847e-645d52a306ee req-d014781a-ca21-4b51-97ff-f122b03e498e service nova] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Updated VIF entry in instance network info cache for port 1fb0e170-03ab-4ee4-9f1f-c5c6008c823d. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 867.549502] env[63088]: DEBUG nova.network.neutron [req-90a8a8cb-f252-43cd-847e-645d52a306ee req-d014781a-ca21-4b51-97ff-f122b03e498e service nova] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Updating instance_info_cache with network_info: [{"id": "1fb0e170-03ab-4ee4-9f1f-c5c6008c823d", "address": "fa:16:3e:a8:02:36", "network": {"id": "dff14a7f-0af1-4e4e-a498-86d7c9816e6b", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-762073671-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f33f2701fad94864a8c406a404bc0a42", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "21310d90-efbc-45a8-a97f-c4358606530f", "external-id": "nsx-vlan-transportzone-672", "segmentation_id": 672, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1fb0e170-03", "ovs_interfaceid": "1fb0e170-03ab-4ee4-9f1f-c5c6008c823d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 867.610034] env[63088]: DEBUG oslo_vmware.api [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285043, 'name': PowerOnVM_Task, 'duration_secs': 0.919111} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.610319] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 9d5f1cde-e787-4a2c-ac49-83ec135ff51c] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 867.610528] env[63088]: INFO nova.compute.manager [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 9d5f1cde-e787-4a2c-ac49-83ec135ff51c] Took 9.09 seconds to spawn the instance on the hypervisor. [ 867.610711] env[63088]: DEBUG nova.compute.manager [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 9d5f1cde-e787-4a2c-ac49-83ec135ff51c] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 867.611550] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5414d707-7061-4aa2-a2e5-2953e3fc8a40 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.657238] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-058a9ba6-53de-4c1c-9276-bca158f57395 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: 58d561fe-26aa-4e94-8d55-cc70c361b479] VM already powered off {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 867.657436] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-058a9ba6-53de-4c1c-9276-bca158f57395 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: 58d561fe-26aa-4e94-8d55-cc70c361b479] Processing image 1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 867.657714] env[63088]: DEBUG oslo_concurrency.lockutils [None req-058a9ba6-53de-4c1c-9276-bca158f57395 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 867.787892] env[63088]: DEBUG oslo_vmware.api [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52ef9333-bd23-1d45-1e8b-037185bbdc23, 'name': SearchDatastore_Task, 'duration_secs': 0.018078} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.788680] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-58a80ef1-fd49-43bb-98c6-7c6677948ad7 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.794018] env[63088]: DEBUG oslo_vmware.api [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Waiting for the task: (returnval){ [ 867.794018] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52d1258d-9bfd-c948-7e32-057da7e3c771" [ 867.794018] env[63088]: _type = "Task" [ 867.794018] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.801789] env[63088]: DEBUG oslo_vmware.api [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52d1258d-9bfd-c948-7e32-057da7e3c771, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.875203] env[63088]: DEBUG oslo_concurrency.lockutils [None req-ad047aea-ed6a-4a24-b3ca-46aa3078b110 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.058s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 867.877454] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.897s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 867.879227] env[63088]: INFO nova.compute.claims [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 0fdb21d1-4111-4ff3-bdc0-e2598298a9a4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 867.888467] env[63088]: DEBUG oslo_vmware.api [None req-ab25717c-7e13-4125-82df-f4c89863e7c3 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285045, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.899111] env[63088]: INFO nova.scheduler.client.report [None req-ad047aea-ed6a-4a24-b3ca-46aa3078b110 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Deleted allocations for instance 07fd3e6f-e0af-467c-9039-238be5c58d25 [ 868.055151] env[63088]: DEBUG oslo_concurrency.lockutils [req-90a8a8cb-f252-43cd-847e-645d52a306ee req-d014781a-ca21-4b51-97ff-f122b03e498e service nova] Releasing lock "refresh_cache-2269e520-08b6-433f-8fe9-7b84d91e02d7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 868.105637] env[63088]: DEBUG nova.network.neutron [req-4a66b9bc-2308-4094-909f-c594cad43562 req-dade105e-15a5-4445-9b6f-e2746d22839b service nova] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Updated VIF entry in instance network info cache for port 81932288-bc32-47c4-a3b2-02597cb6bf81. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 868.106119] env[63088]: DEBUG nova.network.neutron [req-4a66b9bc-2308-4094-909f-c594cad43562 req-dade105e-15a5-4445-9b6f-e2746d22839b service nova] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Updating instance_info_cache with network_info: [{"id": "d5a10951-ae2e-48f5-b0ca-b1f144e5fe3a", "address": "fa:16:3e:95:96:0c", "network": {"id": "1174a284-d3ca-4f9e-aa81-13ee9a693e55", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1994276040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.166", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa3d24a1a6c0430985fd80365d986ee1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd5a10951-ae", "ovs_interfaceid": "d5a10951-ae2e-48f5-b0ca-b1f144e5fe3a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "81932288-bc32-47c4-a3b2-02597cb6bf81", "address": "fa:16:3e:17:d1:1a", "network": {"id": "1174a284-d3ca-4f9e-aa81-13ee9a693e55", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1994276040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa3d24a1a6c0430985fd80365d986ee1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap81932288-bc", "ovs_interfaceid": "81932288-bc32-47c4-a3b2-02597cb6bf81", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 868.128886] env[63088]: INFO nova.compute.manager [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 9d5f1cde-e787-4a2c-ac49-83ec135ff51c] Took 28.87 seconds to build instance. [ 868.305424] env[63088]: DEBUG oslo_vmware.api [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52d1258d-9bfd-c948-7e32-057da7e3c771, 'name': SearchDatastore_Task, 'duration_secs': 0.020245} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.305670] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 868.305939] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore2] 2269e520-08b6-433f-8fe9-7b84d91e02d7/2269e520-08b6-433f-8fe9-7b84d91e02d7.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 868.306284] env[63088]: DEBUG oslo_concurrency.lockutils [None req-058a9ba6-53de-4c1c-9276-bca158f57395 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 868.306506] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-058a9ba6-53de-4c1c-9276-bca158f57395 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 868.306732] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-551a5722-a454-46a5-8803-5cd7c93a1854 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.308738] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ab9ebba7-e39a-44dc-a876-47ec9581220a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.315916] env[63088]: DEBUG oslo_vmware.api [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Waiting for the task: (returnval){ [ 868.315916] env[63088]: value = "task-1285046" [ 868.315916] env[63088]: _type = "Task" [ 868.315916] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.319950] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-058a9ba6-53de-4c1c-9276-bca158f57395 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 868.320152] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-058a9ba6-53de-4c1c-9276-bca158f57395 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 868.321153] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ecd37714-17de-46e6-a6d5-f65ce00e0eee {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.326091] env[63088]: DEBUG oslo_vmware.api [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285046, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.329169] env[63088]: DEBUG oslo_vmware.api [None req-058a9ba6-53de-4c1c-9276-bca158f57395 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Waiting for the task: (returnval){ [ 868.329169] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52a9dffa-00c3-6ab6-f62f-51480b6d51ff" [ 868.329169] env[63088]: _type = "Task" [ 868.329169] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.336765] env[63088]: DEBUG oslo_vmware.api [None req-058a9ba6-53de-4c1c-9276-bca158f57395 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52a9dffa-00c3-6ab6-f62f-51480b6d51ff, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.378121] env[63088]: DEBUG oslo_vmware.api [None req-ab25717c-7e13-4125-82df-f4c89863e7c3 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285045, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.408129] env[63088]: DEBUG oslo_concurrency.lockutils [None req-ad047aea-ed6a-4a24-b3ca-46aa3078b110 tempest-ServerRescueTestJSONUnderV235-484921288 tempest-ServerRescueTestJSONUnderV235-484921288-project-member] Lock "07fd3e6f-e0af-467c-9039-238be5c58d25" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.100s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 868.609823] env[63088]: DEBUG oslo_concurrency.lockutils [req-4a66b9bc-2308-4094-909f-c594cad43562 req-dade105e-15a5-4445-9b6f-e2746d22839b service nova] Releasing lock "refresh_cache-4971b24c-6710-4f50-9846-727dad264b1f" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 868.631637] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d1d29705-fb4f-413f-9196-2d995369af7f tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "9d5f1cde-e787-4a2c-ac49-83ec135ff51c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 33.898s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 868.826689] env[63088]: DEBUG oslo_vmware.api [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285046, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.841376] env[63088]: DEBUG oslo_vmware.api [None req-058a9ba6-53de-4c1c-9276-bca158f57395 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52a9dffa-00c3-6ab6-f62f-51480b6d51ff, 'name': SearchDatastore_Task, 'duration_secs': 0.010186} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.842646] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a09e4783-8987-42c6-bef3-b9240ed515af {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.848851] env[63088]: DEBUG oslo_vmware.api [None req-058a9ba6-53de-4c1c-9276-bca158f57395 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Waiting for the task: (returnval){ [ 868.848851] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]526559a0-1bac-1ee2-7481-bb01ec383095" [ 868.848851] env[63088]: _type = "Task" [ 868.848851] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.857814] env[63088]: DEBUG oslo_vmware.api [None req-058a9ba6-53de-4c1c-9276-bca158f57395 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]526559a0-1bac-1ee2-7481-bb01ec383095, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.878611] env[63088]: DEBUG oslo_vmware.api [None req-ab25717c-7e13-4125-82df-f4c89863e7c3 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285045, 'name': ReconfigVM_Task, 'duration_secs': 1.44657} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.879169] env[63088]: DEBUG oslo_concurrency.lockutils [None req-ab25717c-7e13-4125-82df-f4c89863e7c3 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Releasing lock "4971b24c-6710-4f50-9846-727dad264b1f" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 868.879430] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-ab25717c-7e13-4125-82df-f4c89863e7c3 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Reconfigured VM to attach interface {{(pid=63088) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 868.994575] env[63088]: DEBUG oslo_vmware.rw_handles [None req-0bd62baa-37b1-44d3-beee-8ebfe632ec9b tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/520bd55d-9d5e-31c0-7922-68afdf00ff83/disk-0.vmdk. {{(pid=63088) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 868.995573] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8002b922-02cd-463a-9e06-fa5472f6e7ba {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.002914] env[63088]: DEBUG oslo_vmware.rw_handles [None req-0bd62baa-37b1-44d3-beee-8ebfe632ec9b tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/520bd55d-9d5e-31c0-7922-68afdf00ff83/disk-0.vmdk is in state: ready. {{(pid=63088) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 869.004209] env[63088]: ERROR oslo_vmware.rw_handles [None req-0bd62baa-37b1-44d3-beee-8ebfe632ec9b tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/520bd55d-9d5e-31c0-7922-68afdf00ff83/disk-0.vmdk due to incomplete transfer. [ 869.007328] env[63088]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-4a83ff96-25db-42c4-b530-403a819bd64b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.021363] env[63088]: DEBUG oslo_vmware.rw_handles [None req-0bd62baa-37b1-44d3-beee-8ebfe632ec9b tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/520bd55d-9d5e-31c0-7922-68afdf00ff83/disk-0.vmdk. {{(pid=63088) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 869.021590] env[63088]: DEBUG nova.virt.vmwareapi.images [None req-0bd62baa-37b1-44d3-beee-8ebfe632ec9b tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 8b564e29-96cf-4abf-963d-142b413fb464] Uploaded image f1a9d8b0-fef1-49b9-8a87-99a7556efe13 to the Glance image server {{(pid=63088) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 869.024467] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-0bd62baa-37b1-44d3-beee-8ebfe632ec9b tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 8b564e29-96cf-4abf-963d-142b413fb464] Destroying the VM {{(pid=63088) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 869.028846] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-048b3f3f-d8b5-479d-b920-b53fc0b79831 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.036085] env[63088]: DEBUG oslo_vmware.api [None req-0bd62baa-37b1-44d3-beee-8ebfe632ec9b tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Waiting for the task: (returnval){ [ 869.036085] env[63088]: value = "task-1285047" [ 869.036085] env[63088]: _type = "Task" [ 869.036085] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.045217] env[63088]: DEBUG oslo_vmware.api [None req-0bd62baa-37b1-44d3-beee-8ebfe632ec9b tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285047, 'name': Destroy_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.203931] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f825637d-4542-451c-93bd-408e3c2e72dd {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.212661] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ad0e662-77c5-42e7-a46d-ab04dce716f8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.247808] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e937f694-5bc8-4834-86df-9c256909f4ee {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.255984] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e91a0477-2b92-46af-b4a5-8d734c981ee2 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.271461] env[63088]: DEBUG nova.compute.provider_tree [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 869.326135] env[63088]: DEBUG oslo_vmware.api [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285046, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.524177} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.326427] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore2] 2269e520-08b6-433f-8fe9-7b84d91e02d7/2269e520-08b6-433f-8fe9-7b84d91e02d7.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 869.326646] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Extending root virtual disk to 1048576 {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 869.326923] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4ba11fdf-7dad-4a9d-81cb-6a187496ac3c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.332987] env[63088]: DEBUG oslo_vmware.api [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Waiting for the task: (returnval){ [ 869.332987] env[63088]: value = "task-1285048" [ 869.332987] env[63088]: _type = "Task" [ 869.332987] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.341419] env[63088]: DEBUG oslo_vmware.api [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285048, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.359103] env[63088]: DEBUG oslo_vmware.api [None req-058a9ba6-53de-4c1c-9276-bca158f57395 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]526559a0-1bac-1ee2-7481-bb01ec383095, 'name': SearchDatastore_Task, 'duration_secs': 0.012756} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.359519] env[63088]: DEBUG oslo_concurrency.lockutils [None req-058a9ba6-53de-4c1c-9276-bca158f57395 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 869.359630] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-058a9ba6-53de-4c1c-9276-bca158f57395 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore2] 58d561fe-26aa-4e94-8d55-cc70c361b479/1e8c5d18-0a03-4e18-afe1-de5a6e255953-rescue.vmdk. {{(pid=63088) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 869.359957] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d128f995-417d-403e-ac11-b7d883474970 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.369731] env[63088]: DEBUG oslo_vmware.api [None req-058a9ba6-53de-4c1c-9276-bca158f57395 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Waiting for the task: (returnval){ [ 869.369731] env[63088]: value = "task-1285049" [ 869.369731] env[63088]: _type = "Task" [ 869.369731] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.378165] env[63088]: DEBUG oslo_vmware.api [None req-058a9ba6-53de-4c1c-9276-bca158f57395 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': task-1285049, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.385358] env[63088]: DEBUG oslo_concurrency.lockutils [None req-ab25717c-7e13-4125-82df-f4c89863e7c3 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Lock "interface-4971b24c-6710-4f50-9846-727dad264b1f-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.596s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.548035] env[63088]: DEBUG oslo_vmware.api [None req-0bd62baa-37b1-44d3-beee-8ebfe632ec9b tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285047, 'name': Destroy_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.775966] env[63088]: DEBUG nova.scheduler.client.report [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 869.843716] env[63088]: DEBUG oslo_vmware.api [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285048, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.208766} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.844080] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Extended root virtual disk {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 869.844915] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d0d46d7-7e70-4ad9-a71f-2d892adb0249 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.869907] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Reconfiguring VM instance instance-00000047 to attach disk [datastore2] 2269e520-08b6-433f-8fe9-7b84d91e02d7/2269e520-08b6-433f-8fe9-7b84d91e02d7.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 869.870673] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cdf511e6-b730-4964-8776-ddc31b08ab69 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.904702] env[63088]: DEBUG oslo_vmware.api [None req-058a9ba6-53de-4c1c-9276-bca158f57395 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': task-1285049, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.905217] env[63088]: DEBUG oslo_vmware.api [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Waiting for the task: (returnval){ [ 869.905217] env[63088]: value = "task-1285050" [ 869.905217] env[63088]: _type = "Task" [ 869.905217] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.920127] env[63088]: DEBUG oslo_vmware.api [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285050, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.046919] env[63088]: DEBUG oslo_vmware.api [None req-0bd62baa-37b1-44d3-beee-8ebfe632ec9b tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285047, 'name': Destroy_Task, 'duration_secs': 0.902511} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.047252] env[63088]: INFO nova.virt.vmwareapi.vm_util [None req-0bd62baa-37b1-44d3-beee-8ebfe632ec9b tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 8b564e29-96cf-4abf-963d-142b413fb464] Destroyed the VM [ 870.048144] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-0bd62baa-37b1-44d3-beee-8ebfe632ec9b tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 8b564e29-96cf-4abf-963d-142b413fb464] Deleting Snapshot of the VM instance {{(pid=63088) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 870.048439] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-785af3cf-e136-4cc3-9713-e906c8fb5659 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.054847] env[63088]: DEBUG oslo_vmware.api [None req-0bd62baa-37b1-44d3-beee-8ebfe632ec9b tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Waiting for the task: (returnval){ [ 870.054847] env[63088]: value = "task-1285051" [ 870.054847] env[63088]: _type = "Task" [ 870.054847] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.063394] env[63088]: DEBUG oslo_vmware.api [None req-0bd62baa-37b1-44d3-beee-8ebfe632ec9b tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285051, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.112667] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4ea5fd7b-e84b-41ba-9be2-088f6a9cab78 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Acquiring lock "8b564e29-96cf-4abf-963d-142b413fb464" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.113798] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4ea5fd7b-e84b-41ba-9be2-088f6a9cab78 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Lock "8b564e29-96cf-4abf-963d-142b413fb464" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 870.113798] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4ea5fd7b-e84b-41ba-9be2-088f6a9cab78 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Acquiring lock "8b564e29-96cf-4abf-963d-142b413fb464-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.113798] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4ea5fd7b-e84b-41ba-9be2-088f6a9cab78 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Lock "8b564e29-96cf-4abf-963d-142b413fb464-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 870.113798] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4ea5fd7b-e84b-41ba-9be2-088f6a9cab78 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Lock "8b564e29-96cf-4abf-963d-142b413fb464-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.116037] env[63088]: INFO nova.compute.manager [None req-4ea5fd7b-e84b-41ba-9be2-088f6a9cab78 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 8b564e29-96cf-4abf-963d-142b413fb464] Terminating instance [ 870.118218] env[63088]: DEBUG nova.compute.manager [None req-4ea5fd7b-e84b-41ba-9be2-088f6a9cab78 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 8b564e29-96cf-4abf-963d-142b413fb464] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 870.118463] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-4ea5fd7b-e84b-41ba-9be2-088f6a9cab78 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 8b564e29-96cf-4abf-963d-142b413fb464] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 870.119398] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a563786-b041-48ab-9f2f-b23ffc57af11 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.127668] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-4ea5fd7b-e84b-41ba-9be2-088f6a9cab78 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 8b564e29-96cf-4abf-963d-142b413fb464] Unregistering the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 870.127828] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4562a0b9-96b7-4c07-9a0d-3dffd3acc4d3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.283118] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.406s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.283663] env[63088]: DEBUG nova.compute.manager [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 0fdb21d1-4111-4ff3-bdc0-e2598298a9a4] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 870.286499] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c39e0557-36eb-432c-8e51-8136da99476a tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.255s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 870.286736] env[63088]: DEBUG nova.objects.instance [None req-c39e0557-36eb-432c-8e51-8136da99476a tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Lazy-loading 'resources' on Instance uuid 1d421bf0-925d-408c-a612-b659da92799f {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 870.296361] env[63088]: DEBUG oslo_concurrency.lockutils [None req-24565c48-2c1c-48cc-9d73-37bbd15bd785 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquiring lock "9d5f1cde-e787-4a2c-ac49-83ec135ff51c" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.296635] env[63088]: DEBUG oslo_concurrency.lockutils [None req-24565c48-2c1c-48cc-9d73-37bbd15bd785 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "9d5f1cde-e787-4a2c-ac49-83ec135ff51c" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 870.382901] env[63088]: DEBUG oslo_vmware.api [None req-058a9ba6-53de-4c1c-9276-bca158f57395 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': task-1285049, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.563123} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.383200] env[63088]: INFO nova.virt.vmwareapi.ds_util [None req-058a9ba6-53de-4c1c-9276-bca158f57395 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore2] 58d561fe-26aa-4e94-8d55-cc70c361b479/1e8c5d18-0a03-4e18-afe1-de5a6e255953-rescue.vmdk. [ 870.383965] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b71fd3e-d4a8-46af-873f-6c17ff99306a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.408640] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-058a9ba6-53de-4c1c-9276-bca158f57395 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: 58d561fe-26aa-4e94-8d55-cc70c361b479] Reconfiguring VM instance instance-00000045 to attach disk [datastore2] 58d561fe-26aa-4e94-8d55-cc70c361b479/1e8c5d18-0a03-4e18-afe1-de5a6e255953-rescue.vmdk or device None with type thin {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 870.409048] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-18837eff-f835-4904-9509-d95329b317b9 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.430911] env[63088]: DEBUG oslo_vmware.api [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285050, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.432172] env[63088]: DEBUG oslo_vmware.api [None req-058a9ba6-53de-4c1c-9276-bca158f57395 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Waiting for the task: (returnval){ [ 870.432172] env[63088]: value = "task-1285053" [ 870.432172] env[63088]: _type = "Task" [ 870.432172] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.439544] env[63088]: DEBUG oslo_vmware.api [None req-058a9ba6-53de-4c1c-9276-bca158f57395 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': task-1285053, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.565614] env[63088]: DEBUG oslo_vmware.api [None req-0bd62baa-37b1-44d3-beee-8ebfe632ec9b tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285051, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.700314] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Acquiring lock "254db932-35f1-42e3-9207-cd886efd65a3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.700634] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Lock "254db932-35f1-42e3-9207-cd886efd65a3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 870.789852] env[63088]: DEBUG nova.compute.utils [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 870.794616] env[63088]: DEBUG nova.compute.manager [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 0fdb21d1-4111-4ff3-bdc0-e2598298a9a4] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 870.798423] env[63088]: DEBUG nova.network.neutron [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 0fdb21d1-4111-4ff3-bdc0-e2598298a9a4] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 870.803959] env[63088]: DEBUG nova.compute.utils [None req-24565c48-2c1c-48cc-9d73-37bbd15bd785 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 870.884288] env[63088]: DEBUG nova.policy [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '32d5ff9b3ea345e4ac56010aa917e51b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '03e20d606e654362acbe2b36fe499ae3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 870.921447] env[63088]: DEBUG oslo_vmware.api [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285050, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.944019] env[63088]: DEBUG oslo_vmware.api [None req-058a9ba6-53de-4c1c-9276-bca158f57395 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': task-1285053, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.066857] env[63088]: DEBUG oslo_vmware.api [None req-0bd62baa-37b1-44d3-beee-8ebfe632ec9b tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285051, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.126818] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2863086-3c0c-48ce-a214-00dbd12dab5e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.134035] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51523399-4936-4c54-ace9-00527a0d79ed {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.165541] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-123b742f-c160-4708-80e5-135fe37cb818 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.173385] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-396f70bf-299f-476f-b019-521a6eff9a39 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.186590] env[63088]: DEBUG nova.compute.provider_tree [None req-c39e0557-36eb-432c-8e51-8136da99476a tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Updating inventory in ProviderTree for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 871.203736] env[63088]: DEBUG nova.compute.manager [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 254db932-35f1-42e3-9207-cd886efd65a3] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 871.296553] env[63088]: DEBUG nova.compute.manager [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 0fdb21d1-4111-4ff3-bdc0-e2598298a9a4] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 871.310707] env[63088]: DEBUG oslo_concurrency.lockutils [None req-24565c48-2c1c-48cc-9d73-37bbd15bd785 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "9d5f1cde-e787-4a2c-ac49-83ec135ff51c" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.014s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.350853] env[63088]: DEBUG nova.network.neutron [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 0fdb21d1-4111-4ff3-bdc0-e2598298a9a4] Successfully created port: 94b61c5f-77d5-4f09-9e84-71c70bde1fcb {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 871.421725] env[63088]: DEBUG oslo_vmware.api [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285050, 'name': ReconfigVM_Task, 'duration_secs': 1.289343} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.422097] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Reconfigured VM instance instance-00000047 to attach disk [datastore2] 2269e520-08b6-433f-8fe9-7b84d91e02d7/2269e520-08b6-433f-8fe9-7b84d91e02d7.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 871.422798] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-48684605-6d89-4294-be1d-b8ccfbc7fdfd {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.429516] env[63088]: DEBUG oslo_vmware.api [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Waiting for the task: (returnval){ [ 871.429516] env[63088]: value = "task-1285054" [ 871.429516] env[63088]: _type = "Task" [ 871.429516] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.440143] env[63088]: DEBUG oslo_vmware.api [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285054, 'name': Rename_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.444592] env[63088]: DEBUG oslo_vmware.api [None req-058a9ba6-53de-4c1c-9276-bca158f57395 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': task-1285053, 'name': ReconfigVM_Task, 'duration_secs': 0.578971} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.444851] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-058a9ba6-53de-4c1c-9276-bca158f57395 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: 58d561fe-26aa-4e94-8d55-cc70c361b479] Reconfigured VM instance instance-00000045 to attach disk [datastore2] 58d561fe-26aa-4e94-8d55-cc70c361b479/1e8c5d18-0a03-4e18-afe1-de5a6e255953-rescue.vmdk or device None with type thin {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 871.445922] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2c5d08b-65c3-4103-8f52-7dd7cea3f407 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.471743] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-45d1eae3-6916-4763-a1a8-5ac223206da9 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.486768] env[63088]: DEBUG oslo_vmware.api [None req-058a9ba6-53de-4c1c-9276-bca158f57395 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Waiting for the task: (returnval){ [ 871.486768] env[63088]: value = "task-1285055" [ 871.486768] env[63088]: _type = "Task" [ 871.486768] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.494733] env[63088]: DEBUG oslo_vmware.api [None req-058a9ba6-53de-4c1c-9276-bca158f57395 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': task-1285055, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.516036] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-4ea5fd7b-e84b-41ba-9be2-088f6a9cab78 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 8b564e29-96cf-4abf-963d-142b413fb464] Unregistered the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 871.516411] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-4ea5fd7b-e84b-41ba-9be2-088f6a9cab78 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 8b564e29-96cf-4abf-963d-142b413fb464] Deleting contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 871.516690] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-4ea5fd7b-e84b-41ba-9be2-088f6a9cab78 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Deleting the datastore file [datastore1] 8b564e29-96cf-4abf-963d-142b413fb464 {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 871.517028] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-28d47b7b-caf7-453b-9934-07110a56206c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.523738] env[63088]: DEBUG oslo_vmware.api [None req-4ea5fd7b-e84b-41ba-9be2-088f6a9cab78 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Waiting for the task: (returnval){ [ 871.523738] env[63088]: value = "task-1285056" [ 871.523738] env[63088]: _type = "Task" [ 871.523738] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.531851] env[63088]: DEBUG oslo_vmware.api [None req-4ea5fd7b-e84b-41ba-9be2-088f6a9cab78 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285056, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.567953] env[63088]: DEBUG oslo_vmware.api [None req-0bd62baa-37b1-44d3-beee-8ebfe632ec9b tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285051, 'name': RemoveSnapshot_Task, 'duration_secs': 1.066507} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.570141] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-0bd62baa-37b1-44d3-beee-8ebfe632ec9b tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 8b564e29-96cf-4abf-963d-142b413fb464] Deleted Snapshot of the VM instance {{(pid=63088) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 871.570494] env[63088]: INFO nova.compute.manager [None req-0bd62baa-37b1-44d3-beee-8ebfe632ec9b tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 8b564e29-96cf-4abf-963d-142b413fb464] Took 15.37 seconds to snapshot the instance on the hypervisor. [ 871.598334] env[63088]: DEBUG oslo_concurrency.lockutils [None req-51a6c6bb-2c09-4c64-956d-a4e00c94e93e tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquiring lock "interface-4971b24c-6710-4f50-9846-727dad264b1f-592cc24a-92a9-4bd3-bb9a-a0877ac0db04" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.598512] env[63088]: DEBUG oslo_concurrency.lockutils [None req-51a6c6bb-2c09-4c64-956d-a4e00c94e93e tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Lock "interface-4971b24c-6710-4f50-9846-727dad264b1f-592cc24a-92a9-4bd3-bb9a-a0877ac0db04" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.598902] env[63088]: DEBUG nova.objects.instance [None req-51a6c6bb-2c09-4c64-956d-a4e00c94e93e tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Lazy-loading 'flavor' on Instance uuid 4971b24c-6710-4f50-9846-727dad264b1f {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 871.708370] env[63088]: ERROR nova.scheduler.client.report [None req-c39e0557-36eb-432c-8e51-8136da99476a tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] [req-013ad13f-9b72-4873-8a08-a856b63f3311] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 6eae54a9-8831-40eb-bf54-4bc60d346b02. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-013ad13f-9b72-4873-8a08-a856b63f3311"}]} [ 871.728407] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.730505] env[63088]: DEBUG nova.scheduler.client.report [None req-c39e0557-36eb-432c-8e51-8136da99476a tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Refreshing inventories for resource provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 871.743476] env[63088]: DEBUG nova.scheduler.client.report [None req-c39e0557-36eb-432c-8e51-8136da99476a tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Updating ProviderTree inventory for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 871.743708] env[63088]: DEBUG nova.compute.provider_tree [None req-c39e0557-36eb-432c-8e51-8136da99476a tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Updating inventory in ProviderTree for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 871.759366] env[63088]: DEBUG nova.scheduler.client.report [None req-c39e0557-36eb-432c-8e51-8136da99476a tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Refreshing aggregate associations for resource provider 6eae54a9-8831-40eb-bf54-4bc60d346b02, aggregates: None {{(pid=63088) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 871.776601] env[63088]: DEBUG nova.scheduler.client.report [None req-c39e0557-36eb-432c-8e51-8136da99476a tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Refreshing trait associations for resource provider 6eae54a9-8831-40eb-bf54-4bc60d346b02, traits: COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO {{(pid=63088) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 871.941509] env[63088]: DEBUG oslo_vmware.api [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285054, 'name': Rename_Task, 'duration_secs': 0.148341} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.942405] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 871.942733] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2e8470ae-02df-4786-b002-a19330628f48 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.953472] env[63088]: DEBUG oslo_vmware.api [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Waiting for the task: (returnval){ [ 871.953472] env[63088]: value = "task-1285057" [ 871.953472] env[63088]: _type = "Task" [ 871.953472] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.962252] env[63088]: DEBUG oslo_vmware.api [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285057, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.998397] env[63088]: DEBUG oslo_vmware.api [None req-058a9ba6-53de-4c1c-9276-bca158f57395 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': task-1285055, 'name': ReconfigVM_Task, 'duration_secs': 0.185418} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.001434] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-058a9ba6-53de-4c1c-9276-bca158f57395 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: 58d561fe-26aa-4e94-8d55-cc70c361b479] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 872.003010] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-61d300b6-331d-43ab-848c-03805a9ac399 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.008732] env[63088]: DEBUG oslo_vmware.api [None req-058a9ba6-53de-4c1c-9276-bca158f57395 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Waiting for the task: (returnval){ [ 872.008732] env[63088]: value = "task-1285058" [ 872.008732] env[63088]: _type = "Task" [ 872.008732] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.016576] env[63088]: DEBUG oslo_vmware.api [None req-058a9ba6-53de-4c1c-9276-bca158f57395 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': task-1285058, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.032557] env[63088]: DEBUG oslo_vmware.api [None req-4ea5fd7b-e84b-41ba-9be2-088f6a9cab78 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285056, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.150478} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.035638] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-4ea5fd7b-e84b-41ba-9be2-088f6a9cab78 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 872.035638] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-4ea5fd7b-e84b-41ba-9be2-088f6a9cab78 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 8b564e29-96cf-4abf-963d-142b413fb464] Deleted contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 872.035638] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-4ea5fd7b-e84b-41ba-9be2-088f6a9cab78 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 8b564e29-96cf-4abf-963d-142b413fb464] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 872.035800] env[63088]: INFO nova.compute.manager [None req-4ea5fd7b-e84b-41ba-9be2-088f6a9cab78 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 8b564e29-96cf-4abf-963d-142b413fb464] Took 1.92 seconds to destroy the instance on the hypervisor. [ 872.036026] env[63088]: DEBUG oslo.service.loopingcall [None req-4ea5fd7b-e84b-41ba-9be2-088f6a9cab78 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 872.036617] env[63088]: DEBUG nova.compute.manager [-] [instance: 8b564e29-96cf-4abf-963d-142b413fb464] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 872.036738] env[63088]: DEBUG nova.network.neutron [-] [instance: 8b564e29-96cf-4abf-963d-142b413fb464] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 872.077196] env[63088]: DEBUG nova.compute.manager [None req-0bd62baa-37b1-44d3-beee-8ebfe632ec9b tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 8b564e29-96cf-4abf-963d-142b413fb464] Instance disappeared during snapshot {{(pid=63088) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4494}} [ 872.094285] env[63088]: DEBUG nova.compute.manager [None req-0bd62baa-37b1-44d3-beee-8ebfe632ec9b tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Image not found during clean up f1a9d8b0-fef1-49b9-8a87-99a7556efe13 {{(pid=63088) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4500}} [ 872.099366] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9448ec25-40b0-423d-948b-9cc8b38b6c1c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.115196] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a3038aa-4fbd-4c1d-82ad-c30841c48b1e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.153356] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afa72f5f-b746-4662-a99a-f35100a94e52 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.161396] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ad7b7c8-6af5-4233-a5db-8c583b277a2e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.176408] env[63088]: DEBUG nova.compute.provider_tree [None req-c39e0557-36eb-432c-8e51-8136da99476a tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 872.238115] env[63088]: DEBUG nova.objects.instance [None req-51a6c6bb-2c09-4c64-956d-a4e00c94e93e tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Lazy-loading 'pci_requests' on Instance uuid 4971b24c-6710-4f50-9846-727dad264b1f {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 872.304523] env[63088]: DEBUG nova.compute.manager [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 0fdb21d1-4111-4ff3-bdc0-e2598298a9a4] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 872.330506] env[63088]: DEBUG nova.virt.hardware [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 872.330763] env[63088]: DEBUG nova.virt.hardware [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 872.330942] env[63088]: DEBUG nova.virt.hardware [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 872.332476] env[63088]: DEBUG nova.virt.hardware [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 872.332589] env[63088]: DEBUG nova.virt.hardware [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 872.332788] env[63088]: DEBUG nova.virt.hardware [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 872.332937] env[63088]: DEBUG nova.virt.hardware [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 872.333190] env[63088]: DEBUG nova.virt.hardware [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 872.333385] env[63088]: DEBUG nova.virt.hardware [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 872.333567] env[63088]: DEBUG nova.virt.hardware [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 872.333920] env[63088]: DEBUG nova.virt.hardware [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 872.334803] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f24720e-6d96-475d-9ee4-42528d77d9a2 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.344959] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01ea6a59-5bfd-4602-8661-c62f66475930 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.401183] env[63088]: DEBUG oslo_concurrency.lockutils [None req-24565c48-2c1c-48cc-9d73-37bbd15bd785 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquiring lock "9d5f1cde-e787-4a2c-ac49-83ec135ff51c" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.401473] env[63088]: DEBUG oslo_concurrency.lockutils [None req-24565c48-2c1c-48cc-9d73-37bbd15bd785 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "9d5f1cde-e787-4a2c-ac49-83ec135ff51c" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.401752] env[63088]: INFO nova.compute.manager [None req-24565c48-2c1c-48cc-9d73-37bbd15bd785 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 9d5f1cde-e787-4a2c-ac49-83ec135ff51c] Attaching volume b674aa10-140a-4cb0-8ca0-7fa9fddf7a6d to /dev/sdb [ 872.440564] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24d78e05-b61c-4165-b92a-db5b490ff765 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.448549] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92cf3097-6d62-42fd-bcba-b46ec122210a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.462110] env[63088]: DEBUG nova.virt.block_device [None req-24565c48-2c1c-48cc-9d73-37bbd15bd785 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 9d5f1cde-e787-4a2c-ac49-83ec135ff51c] Updating existing volume attachment record: 75cd854e-d98f-40d2-9d8c-35f4aa6dd713 {{(pid=63088) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 872.467400] env[63088]: DEBUG oslo_vmware.api [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285057, 'name': PowerOnVM_Task, 'duration_secs': 0.467672} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.467657] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 872.467864] env[63088]: INFO nova.compute.manager [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Took 8.45 seconds to spawn the instance on the hypervisor. [ 872.468055] env[63088]: DEBUG nova.compute.manager [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 872.468823] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acd8b6a9-2019-466c-ad0b-5211b2079626 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.519804] env[63088]: DEBUG oslo_vmware.api [None req-058a9ba6-53de-4c1c-9276-bca158f57395 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': task-1285058, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.679284] env[63088]: DEBUG nova.scheduler.client.report [None req-c39e0557-36eb-432c-8e51-8136da99476a tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 872.741373] env[63088]: DEBUG nova.objects.base [None req-51a6c6bb-2c09-4c64-956d-a4e00c94e93e tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Object Instance<4971b24c-6710-4f50-9846-727dad264b1f> lazy-loaded attributes: flavor,pci_requests {{(pid=63088) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 872.741699] env[63088]: DEBUG nova.network.neutron [None req-51a6c6bb-2c09-4c64-956d-a4e00c94e93e tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 872.919346] env[63088]: DEBUG nova.policy [None req-51a6c6bb-2c09-4c64-956d-a4e00c94e93e tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '93f1d334e71a4a99ba300c3ee6d70243', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'aa3d24a1a6c0430985fd80365d986ee1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 872.988703] env[63088]: DEBUG nova.network.neutron [-] [instance: 8b564e29-96cf-4abf-963d-142b413fb464] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 872.996949] env[63088]: INFO nova.compute.manager [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Took 30.07 seconds to build instance. [ 873.024154] env[63088]: DEBUG oslo_vmware.api [None req-058a9ba6-53de-4c1c-9276-bca158f57395 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': task-1285058, 'name': PowerOnVM_Task, 'duration_secs': 0.745831} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.024154] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-058a9ba6-53de-4c1c-9276-bca158f57395 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: 58d561fe-26aa-4e94-8d55-cc70c361b479] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 873.029606] env[63088]: DEBUG nova.compute.manager [None req-058a9ba6-53de-4c1c-9276-bca158f57395 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: 58d561fe-26aa-4e94-8d55-cc70c361b479] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 873.030916] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f50f078b-a29f-4f4c-a7c0-211452745224 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.184271] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c39e0557-36eb-432c-8e51-8136da99476a tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.898s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 873.186662] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f74cf871-55f9-46f3-b9cb-4b34c2963a40 tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.145s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 873.186903] env[63088]: DEBUG nova.objects.instance [None req-f74cf871-55f9-46f3-b9cb-4b34c2963a40 tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Lazy-loading 'resources' on Instance uuid 00925f94-dbf9-453c-a124-d8434679aedf {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 873.216977] env[63088]: INFO nova.scheduler.client.report [None req-c39e0557-36eb-432c-8e51-8136da99476a tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Deleted allocations for instance 1d421bf0-925d-408c-a612-b659da92799f [ 873.352086] env[63088]: DEBUG nova.compute.manager [req-cc31017e-00c6-484f-ae99-4a0af8d0dac6 req-4ba164ae-b0c2-439a-9342-a89348dd70e3 service nova] [instance: 8b564e29-96cf-4abf-963d-142b413fb464] Received event network-vif-deleted-7a0cbd40-cac8-4bb6-944c-486989457770 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 873.497584] env[63088]: INFO nova.compute.manager [-] [instance: 8b564e29-96cf-4abf-963d-142b413fb464] Took 1.46 seconds to deallocate network for instance. [ 873.499725] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3dc9a54c-7710-4a02-8e4a-c8a6a8888f99 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Lock "2269e520-08b6-433f-8fe9-7b84d91e02d7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 36.030s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 873.726232] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c39e0557-36eb-432c-8e51-8136da99476a tempest-VolumesAdminNegativeTest-719871167 tempest-VolumesAdminNegativeTest-719871167-project-member] Lock "1d421bf0-925d-408c-a612-b659da92799f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.002s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 873.784639] env[63088]: DEBUG nova.network.neutron [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 0fdb21d1-4111-4ff3-bdc0-e2598298a9a4] Successfully updated port: 94b61c5f-77d5-4f09-9e84-71c70bde1fcb {{(pid=63088) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 873.963022] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6de47f7-ac1d-443b-ada9-11ad239df3c6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.972498] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0b3ce5f-16b5-425d-a92c-608a513645b4 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.011208] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4ea5fd7b-e84b-41ba-9be2-088f6a9cab78 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.012345] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56d2d5af-80ab-4898-9cfc-f8826b62c06c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.021885] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d32c3a74-d363-488e-8ff9-f7d31985178b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.037679] env[63088]: DEBUG nova.compute.provider_tree [None req-f74cf871-55f9-46f3-b9cb-4b34c2963a40 tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 874.287130] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquiring lock "refresh_cache-0fdb21d1-4111-4ff3-bdc0-e2598298a9a4" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 874.287284] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquired lock "refresh_cache-0fdb21d1-4111-4ff3-bdc0-e2598298a9a4" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 874.287439] env[63088]: DEBUG nova.network.neutron [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 0fdb21d1-4111-4ff3-bdc0-e2598298a9a4] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 874.546745] env[63088]: DEBUG nova.scheduler.client.report [None req-f74cf871-55f9-46f3-b9cb-4b34c2963a40 tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 874.871089] env[63088]: DEBUG nova.network.neutron [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 0fdb21d1-4111-4ff3-bdc0-e2598298a9a4] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 874.908260] env[63088]: INFO nova.compute.manager [None req-6c4b10e7-44a6-41b3-b172-43d52124d161 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: 58d561fe-26aa-4e94-8d55-cc70c361b479] Unrescuing [ 874.908641] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6c4b10e7-44a6-41b3-b172-43d52124d161 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Acquiring lock "refresh_cache-58d561fe-26aa-4e94-8d55-cc70c361b479" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 874.909083] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6c4b10e7-44a6-41b3-b172-43d52124d161 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Acquired lock "refresh_cache-58d561fe-26aa-4e94-8d55-cc70c361b479" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 874.909243] env[63088]: DEBUG nova.network.neutron [None req-6c4b10e7-44a6-41b3-b172-43d52124d161 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: 58d561fe-26aa-4e94-8d55-cc70c361b479] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 875.057096] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f74cf871-55f9-46f3-b9cb-4b34c2963a40 tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.870s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.063819] env[63088]: DEBUG oslo_concurrency.lockutils [None req-2790830b-05cd-47fb-bd8c-69ead0f38b2f tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.929s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.064253] env[63088]: DEBUG nova.objects.instance [None req-2790830b-05cd-47fb-bd8c-69ead0f38b2f tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Lazy-loading 'resources' on Instance uuid 7f341d80-450a-47b8-b26d-15f2ce3e378c {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 875.089287] env[63088]: INFO nova.scheduler.client.report [None req-f74cf871-55f9-46f3-b9cb-4b34c2963a40 tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Deleted allocations for instance 00925f94-dbf9-453c-a124-d8434679aedf [ 875.282258] env[63088]: DEBUG nova.network.neutron [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 0fdb21d1-4111-4ff3-bdc0-e2598298a9a4] Updating instance_info_cache with network_info: [{"id": "94b61c5f-77d5-4f09-9e84-71c70bde1fcb", "address": "fa:16:3e:7b:c3:4b", "network": {"id": "307b966a-d9e1-40fd-9313-1ad94c734308", "bridge": "br-int", "label": "tempest-ServersTestJSON-1907249197-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "03e20d606e654362acbe2b36fe499ae3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae18b41f-e73c-44f1-83dd-467c080944f4", "external-id": "nsx-vlan-transportzone-653", "segmentation_id": 653, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap94b61c5f-77", "ovs_interfaceid": "94b61c5f-77d5-4f09-9e84-71c70bde1fcb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 875.518268] env[63088]: DEBUG nova.compute.manager [req-60895767-8856-47f8-8523-422d1647aa2d req-c7d4ecfc-3b08-4118-b674-34a35a24fdcc service nova] [instance: 0fdb21d1-4111-4ff3-bdc0-e2598298a9a4] Received event network-vif-plugged-94b61c5f-77d5-4f09-9e84-71c70bde1fcb {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 875.518268] env[63088]: DEBUG oslo_concurrency.lockutils [req-60895767-8856-47f8-8523-422d1647aa2d req-c7d4ecfc-3b08-4118-b674-34a35a24fdcc service nova] Acquiring lock "0fdb21d1-4111-4ff3-bdc0-e2598298a9a4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.518268] env[63088]: DEBUG oslo_concurrency.lockutils [req-60895767-8856-47f8-8523-422d1647aa2d req-c7d4ecfc-3b08-4118-b674-34a35a24fdcc service nova] Lock "0fdb21d1-4111-4ff3-bdc0-e2598298a9a4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.518268] env[63088]: DEBUG oslo_concurrency.lockutils [req-60895767-8856-47f8-8523-422d1647aa2d req-c7d4ecfc-3b08-4118-b674-34a35a24fdcc service nova] Lock "0fdb21d1-4111-4ff3-bdc0-e2598298a9a4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.518268] env[63088]: DEBUG nova.compute.manager [req-60895767-8856-47f8-8523-422d1647aa2d req-c7d4ecfc-3b08-4118-b674-34a35a24fdcc service nova] [instance: 0fdb21d1-4111-4ff3-bdc0-e2598298a9a4] No waiting events found dispatching network-vif-plugged-94b61c5f-77d5-4f09-9e84-71c70bde1fcb {{(pid=63088) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 875.518268] env[63088]: WARNING nova.compute.manager [req-60895767-8856-47f8-8523-422d1647aa2d req-c7d4ecfc-3b08-4118-b674-34a35a24fdcc service nova] [instance: 0fdb21d1-4111-4ff3-bdc0-e2598298a9a4] Received unexpected event network-vif-plugged-94b61c5f-77d5-4f09-9e84-71c70bde1fcb for instance with vm_state building and task_state spawning. [ 875.518268] env[63088]: DEBUG nova.compute.manager [req-60895767-8856-47f8-8523-422d1647aa2d req-c7d4ecfc-3b08-4118-b674-34a35a24fdcc service nova] [instance: 0fdb21d1-4111-4ff3-bdc0-e2598298a9a4] Received event network-changed-94b61c5f-77d5-4f09-9e84-71c70bde1fcb {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 875.521564] env[63088]: DEBUG nova.compute.manager [req-60895767-8856-47f8-8523-422d1647aa2d req-c7d4ecfc-3b08-4118-b674-34a35a24fdcc service nova] [instance: 0fdb21d1-4111-4ff3-bdc0-e2598298a9a4] Refreshing instance network info cache due to event network-changed-94b61c5f-77d5-4f09-9e84-71c70bde1fcb. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 875.521564] env[63088]: DEBUG oslo_concurrency.lockutils [req-60895767-8856-47f8-8523-422d1647aa2d req-c7d4ecfc-3b08-4118-b674-34a35a24fdcc service nova] Acquiring lock "refresh_cache-0fdb21d1-4111-4ff3-bdc0-e2598298a9a4" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 875.559440] env[63088]: DEBUG nova.network.neutron [None req-51a6c6bb-2c09-4c64-956d-a4e00c94e93e tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Successfully updated port: 592cc24a-92a9-4bd3-bb9a-a0877ac0db04 {{(pid=63088) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 875.605611] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f74cf871-55f9-46f3-b9cb-4b34c2963a40 tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Lock "00925f94-dbf9-453c-a124-d8434679aedf" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.418s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.785403] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Releasing lock "refresh_cache-0fdb21d1-4111-4ff3-bdc0-e2598298a9a4" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 875.785833] env[63088]: DEBUG nova.compute.manager [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 0fdb21d1-4111-4ff3-bdc0-e2598298a9a4] Instance network_info: |[{"id": "94b61c5f-77d5-4f09-9e84-71c70bde1fcb", "address": "fa:16:3e:7b:c3:4b", "network": {"id": "307b966a-d9e1-40fd-9313-1ad94c734308", "bridge": "br-int", "label": "tempest-ServersTestJSON-1907249197-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "03e20d606e654362acbe2b36fe499ae3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae18b41f-e73c-44f1-83dd-467c080944f4", "external-id": "nsx-vlan-transportzone-653", "segmentation_id": 653, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap94b61c5f-77", "ovs_interfaceid": "94b61c5f-77d5-4f09-9e84-71c70bde1fcb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 875.786481] env[63088]: DEBUG oslo_concurrency.lockutils [req-60895767-8856-47f8-8523-422d1647aa2d req-c7d4ecfc-3b08-4118-b674-34a35a24fdcc service nova] Acquired lock "refresh_cache-0fdb21d1-4111-4ff3-bdc0-e2598298a9a4" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 875.786481] env[63088]: DEBUG nova.network.neutron [req-60895767-8856-47f8-8523-422d1647aa2d req-c7d4ecfc-3b08-4118-b674-34a35a24fdcc service nova] [instance: 0fdb21d1-4111-4ff3-bdc0-e2598298a9a4] Refreshing network info cache for port 94b61c5f-77d5-4f09-9e84-71c70bde1fcb {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 875.788194] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 0fdb21d1-4111-4ff3-bdc0-e2598298a9a4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7b:c3:4b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ae18b41f-e73c-44f1-83dd-467c080944f4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '94b61c5f-77d5-4f09-9e84-71c70bde1fcb', 'vif_model': 'vmxnet3'}] {{(pid=63088) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 875.798793] env[63088]: DEBUG oslo.service.loopingcall [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 875.802870] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0fdb21d1-4111-4ff3-bdc0-e2598298a9a4] Creating VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 875.803993] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2249a78d-5223-4967-b2e4-cc900150a8d4 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.828091] env[63088]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 875.828091] env[63088]: value = "task-1285062" [ 875.828091] env[63088]: _type = "Task" [ 875.828091] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.830445] env[63088]: DEBUG nova.network.neutron [None req-6c4b10e7-44a6-41b3-b172-43d52124d161 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: 58d561fe-26aa-4e94-8d55-cc70c361b479] Updating instance_info_cache with network_info: [{"id": "c6477c64-3129-4546-84b1-40bc79bcaa57", "address": "fa:16:3e:75:e7:1a", "network": {"id": "7db9302e-1c7a-42dc-8084-98e2a57ad7bf", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-2073161169-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "900466c3c09a4f928bdd4b602715c838", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "87bbf4e0-9064-4516-b7e7-44973f817205", "external-id": "nsx-vlan-transportzone-507", "segmentation_id": 507, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6477c64-31", "ovs_interfaceid": "c6477c64-3129-4546-84b1-40bc79bcaa57", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 875.843904] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285062, 'name': CreateVM_Task} progress is 6%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.856957] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c7147fb-d170-4977-86fa-7c2111295826 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.864048] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3faf5103-10dc-4778-af49-c19c05d35316 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.901155] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d9e962a-cf84-48e6-bd12-0162728bd0ad {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.910616] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ddaa6df-86c8-457b-9836-9893139541d9 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.926423] env[63088]: DEBUG nova.compute.provider_tree [None req-2790830b-05cd-47fb-bd8c-69ead0f38b2f tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 876.062028] env[63088]: DEBUG oslo_concurrency.lockutils [None req-51a6c6bb-2c09-4c64-956d-a4e00c94e93e tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquiring lock "refresh_cache-4971b24c-6710-4f50-9846-727dad264b1f" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 876.062331] env[63088]: DEBUG oslo_concurrency.lockutils [None req-51a6c6bb-2c09-4c64-956d-a4e00c94e93e tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquired lock "refresh_cache-4971b24c-6710-4f50-9846-727dad264b1f" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.064733] env[63088]: DEBUG nova.network.neutron [None req-51a6c6bb-2c09-4c64-956d-a4e00c94e93e tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 876.332832] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6c4b10e7-44a6-41b3-b172-43d52124d161 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Releasing lock "refresh_cache-58d561fe-26aa-4e94-8d55-cc70c361b479" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 876.333582] env[63088]: DEBUG nova.objects.instance [None req-6c4b10e7-44a6-41b3-b172-43d52124d161 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Lazy-loading 'flavor' on Instance uuid 58d561fe-26aa-4e94-8d55-cc70c361b479 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 876.346749] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285062, 'name': CreateVM_Task, 'duration_secs': 0.394409} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.350206] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0fdb21d1-4111-4ff3-bdc0-e2598298a9a4] Created VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 876.351087] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 876.351244] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.351616] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 876.352088] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5ae157f8-70ca-41ee-befd-ee77b6d287da {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.356649] env[63088]: DEBUG oslo_vmware.api [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Waiting for the task: (returnval){ [ 876.356649] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52252b7d-eecc-f7b1-21a2-08d19af5acee" [ 876.356649] env[63088]: _type = "Task" [ 876.356649] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.360658] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0dea949d-9593-45db-b8f6-0d07193e67be tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Acquiring lock "34ce411f-40c8-446e-b685-cd3ce07663dd" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 876.360877] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0dea949d-9593-45db-b8f6-0d07193e67be tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Lock "34ce411f-40c8-446e-b685-cd3ce07663dd" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 876.361113] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0dea949d-9593-45db-b8f6-0d07193e67be tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Acquiring lock "34ce411f-40c8-446e-b685-cd3ce07663dd-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 876.361309] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0dea949d-9593-45db-b8f6-0d07193e67be tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Lock "34ce411f-40c8-446e-b685-cd3ce07663dd-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 876.361477] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0dea949d-9593-45db-b8f6-0d07193e67be tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Lock "34ce411f-40c8-446e-b685-cd3ce07663dd-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 876.363492] env[63088]: INFO nova.compute.manager [None req-0dea949d-9593-45db-b8f6-0d07193e67be tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] [instance: 34ce411f-40c8-446e-b685-cd3ce07663dd] Terminating instance [ 876.367513] env[63088]: DEBUG oslo_vmware.api [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52252b7d-eecc-f7b1-21a2-08d19af5acee, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.368075] env[63088]: DEBUG nova.compute.manager [None req-0dea949d-9593-45db-b8f6-0d07193e67be tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] [instance: 34ce411f-40c8-446e-b685-cd3ce07663dd] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 876.368273] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-0dea949d-9593-45db-b8f6-0d07193e67be tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] [instance: 34ce411f-40c8-446e-b685-cd3ce07663dd] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 876.368996] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-023f8ec7-e574-44ed-86ef-bb441e276f7e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.376035] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-0dea949d-9593-45db-b8f6-0d07193e67be tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] [instance: 34ce411f-40c8-446e-b685-cd3ce07663dd] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 876.376182] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-52313612-5b1e-4d37-9b60-eba58501d21a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.381906] env[63088]: DEBUG oslo_vmware.api [None req-0dea949d-9593-45db-b8f6-0d07193e67be tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Waiting for the task: (returnval){ [ 876.381906] env[63088]: value = "task-1285063" [ 876.381906] env[63088]: _type = "Task" [ 876.381906] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.391250] env[63088]: DEBUG oslo_vmware.api [None req-0dea949d-9593-45db-b8f6-0d07193e67be tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Task: {'id': task-1285063, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.429114] env[63088]: DEBUG nova.scheduler.client.report [None req-2790830b-05cd-47fb-bd8c-69ead0f38b2f tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 876.564178] env[63088]: DEBUG nova.network.neutron [req-60895767-8856-47f8-8523-422d1647aa2d req-c7d4ecfc-3b08-4118-b674-34a35a24fdcc service nova] [instance: 0fdb21d1-4111-4ff3-bdc0-e2598298a9a4] Updated VIF entry in instance network info cache for port 94b61c5f-77d5-4f09-9e84-71c70bde1fcb. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 876.564178] env[63088]: DEBUG nova.network.neutron [req-60895767-8856-47f8-8523-422d1647aa2d req-c7d4ecfc-3b08-4118-b674-34a35a24fdcc service nova] [instance: 0fdb21d1-4111-4ff3-bdc0-e2598298a9a4] Updating instance_info_cache with network_info: [{"id": "94b61c5f-77d5-4f09-9e84-71c70bde1fcb", "address": "fa:16:3e:7b:c3:4b", "network": {"id": "307b966a-d9e1-40fd-9313-1ad94c734308", "bridge": "br-int", "label": "tempest-ServersTestJSON-1907249197-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "03e20d606e654362acbe2b36fe499ae3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae18b41f-e73c-44f1-83dd-467c080944f4", "external-id": "nsx-vlan-transportzone-653", "segmentation_id": 653, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap94b61c5f-77", "ovs_interfaceid": "94b61c5f-77d5-4f09-9e84-71c70bde1fcb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 876.608081] env[63088]: WARNING nova.network.neutron [None req-51a6c6bb-2c09-4c64-956d-a4e00c94e93e tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] 1174a284-d3ca-4f9e-aa81-13ee9a693e55 already exists in list: networks containing: ['1174a284-d3ca-4f9e-aa81-13ee9a693e55']. ignoring it [ 876.608756] env[63088]: WARNING nova.network.neutron [None req-51a6c6bb-2c09-4c64-956d-a4e00c94e93e tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] 1174a284-d3ca-4f9e-aa81-13ee9a693e55 already exists in list: networks containing: ['1174a284-d3ca-4f9e-aa81-13ee9a693e55']. ignoring it [ 876.705907] env[63088]: DEBUG nova.compute.manager [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Stashing vm_state: active {{(pid=63088) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 876.849995] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44337300-6acf-4c04-8dcc-9a05b0d075c0 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.871874] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c4b10e7-44a6-41b3-b172-43d52124d161 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: 58d561fe-26aa-4e94-8d55-cc70c361b479] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 876.875239] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-08623769-ff91-471a-af3a-a960492fa8d4 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.882411] env[63088]: DEBUG oslo_vmware.api [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52252b7d-eecc-f7b1-21a2-08d19af5acee, 'name': SearchDatastore_Task, 'duration_secs': 0.011276} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.886580] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 876.886580] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 0fdb21d1-4111-4ff3-bdc0-e2598298a9a4] Processing image 1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 876.886580] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 876.886899] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.886963] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 876.890565] env[63088]: DEBUG oslo_vmware.api [None req-6c4b10e7-44a6-41b3-b172-43d52124d161 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Waiting for the task: (returnval){ [ 876.890565] env[63088]: value = "task-1285065" [ 876.890565] env[63088]: _type = "Task" [ 876.890565] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.890818] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-39cd29e6-ae87-429e-b4ae-f251a8e68a6f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.902690] env[63088]: DEBUG oslo_vmware.api [None req-0dea949d-9593-45db-b8f6-0d07193e67be tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Task: {'id': task-1285063, 'name': PowerOffVM_Task, 'duration_secs': 0.45492} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.903090] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-0dea949d-9593-45db-b8f6-0d07193e67be tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] [instance: 34ce411f-40c8-446e-b685-cd3ce07663dd] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 876.903218] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-0dea949d-9593-45db-b8f6-0d07193e67be tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] [instance: 34ce411f-40c8-446e-b685-cd3ce07663dd] Unregistering the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 876.903487] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-23cf550e-a8c8-40a0-8e4e-4af100a69db2 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.908277] env[63088]: DEBUG oslo_vmware.api [None req-6c4b10e7-44a6-41b3-b172-43d52124d161 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': task-1285065, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.909703] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 876.910057] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 876.910630] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-34e94f29-9f51-467b-bf7d-9d514a078624 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.916248] env[63088]: DEBUG oslo_vmware.api [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Waiting for the task: (returnval){ [ 876.916248] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]5203369c-b1d0-a7dd-9995-995f8f5cf729" [ 876.916248] env[63088]: _type = "Task" [ 876.916248] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.925554] env[63088]: DEBUG oslo_vmware.api [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]5203369c-b1d0-a7dd-9995-995f8f5cf729, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.934201] env[63088]: DEBUG oslo_concurrency.lockutils [None req-2790830b-05cd-47fb-bd8c-69ead0f38b2f tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.870s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 876.937771] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.266s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 876.939990] env[63088]: INFO nova.compute.claims [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: 667fd5e9-5fe4-41e1-9d8a-896c6e5c6286] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 876.964786] env[63088]: INFO nova.scheduler.client.report [None req-2790830b-05cd-47fb-bd8c-69ead0f38b2f tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Deleted allocations for instance 7f341d80-450a-47b8-b26d-15f2ce3e378c [ 876.977819] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-0dea949d-9593-45db-b8f6-0d07193e67be tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] [instance: 34ce411f-40c8-446e-b685-cd3ce07663dd] Unregistered the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 876.977819] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-0dea949d-9593-45db-b8f6-0d07193e67be tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] [instance: 34ce411f-40c8-446e-b685-cd3ce07663dd] Deleting contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 876.977930] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-0dea949d-9593-45db-b8f6-0d07193e67be tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Deleting the datastore file [datastore1] 34ce411f-40c8-446e-b685-cd3ce07663dd {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 876.978201] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-074b6e5a-9521-4368-9e34-976d55779706 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.984713] env[63088]: DEBUG oslo_vmware.api [None req-0dea949d-9593-45db-b8f6-0d07193e67be tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Waiting for the task: (returnval){ [ 876.984713] env[63088]: value = "task-1285067" [ 876.984713] env[63088]: _type = "Task" [ 876.984713] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.996248] env[63088]: DEBUG oslo_vmware.api [None req-0dea949d-9593-45db-b8f6-0d07193e67be tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Task: {'id': task-1285067, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.069337] env[63088]: DEBUG oslo_concurrency.lockutils [req-60895767-8856-47f8-8523-422d1647aa2d req-c7d4ecfc-3b08-4118-b674-34a35a24fdcc service nova] Releasing lock "refresh_cache-0fdb21d1-4111-4ff3-bdc0-e2598298a9a4" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.069991] env[63088]: DEBUG nova.compute.manager [req-60895767-8856-47f8-8523-422d1647aa2d req-c7d4ecfc-3b08-4118-b674-34a35a24fdcc service nova] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Received event network-vif-plugged-592cc24a-92a9-4bd3-bb9a-a0877ac0db04 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 877.069991] env[63088]: DEBUG oslo_concurrency.lockutils [req-60895767-8856-47f8-8523-422d1647aa2d req-c7d4ecfc-3b08-4118-b674-34a35a24fdcc service nova] Acquiring lock "4971b24c-6710-4f50-9846-727dad264b1f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 877.070224] env[63088]: DEBUG oslo_concurrency.lockutils [req-60895767-8856-47f8-8523-422d1647aa2d req-c7d4ecfc-3b08-4118-b674-34a35a24fdcc service nova] Lock "4971b24c-6710-4f50-9846-727dad264b1f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.070442] env[63088]: DEBUG oslo_concurrency.lockutils [req-60895767-8856-47f8-8523-422d1647aa2d req-c7d4ecfc-3b08-4118-b674-34a35a24fdcc service nova] Lock "4971b24c-6710-4f50-9846-727dad264b1f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.070716] env[63088]: DEBUG nova.compute.manager [req-60895767-8856-47f8-8523-422d1647aa2d req-c7d4ecfc-3b08-4118-b674-34a35a24fdcc service nova] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] No waiting events found dispatching network-vif-plugged-592cc24a-92a9-4bd3-bb9a-a0877ac0db04 {{(pid=63088) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 877.070938] env[63088]: WARNING nova.compute.manager [req-60895767-8856-47f8-8523-422d1647aa2d req-c7d4ecfc-3b08-4118-b674-34a35a24fdcc service nova] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Received unexpected event network-vif-plugged-592cc24a-92a9-4bd3-bb9a-a0877ac0db04 for instance with vm_state active and task_state None. [ 877.175719] env[63088]: DEBUG oslo_concurrency.lockutils [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Acquiring lock "71e37d8e-a454-46c4-a3cc-3d5671a32beb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 877.175719] env[63088]: DEBUG oslo_concurrency.lockutils [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Lock "71e37d8e-a454-46c4-a3cc-3d5671a32beb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.225763] env[63088]: DEBUG oslo_concurrency.lockutils [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 877.256377] env[63088]: DEBUG nova.network.neutron [None req-51a6c6bb-2c09-4c64-956d-a4e00c94e93e tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Updating instance_info_cache with network_info: [{"id": "d5a10951-ae2e-48f5-b0ca-b1f144e5fe3a", "address": "fa:16:3e:95:96:0c", "network": {"id": "1174a284-d3ca-4f9e-aa81-13ee9a693e55", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1994276040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.166", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa3d24a1a6c0430985fd80365d986ee1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd5a10951-ae", "ovs_interfaceid": "d5a10951-ae2e-48f5-b0ca-b1f144e5fe3a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "81932288-bc32-47c4-a3b2-02597cb6bf81", "address": "fa:16:3e:17:d1:1a", "network": {"id": "1174a284-d3ca-4f9e-aa81-13ee9a693e55", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1994276040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa3d24a1a6c0430985fd80365d986ee1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap81932288-bc", "ovs_interfaceid": "81932288-bc32-47c4-a3b2-02597cb6bf81", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "592cc24a-92a9-4bd3-bb9a-a0877ac0db04", "address": "fa:16:3e:fe:d7:1a", "network": {"id": "1174a284-d3ca-4f9e-aa81-13ee9a693e55", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1994276040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa3d24a1a6c0430985fd80365d986ee1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap592cc24a-92", "ovs_interfaceid": "592cc24a-92a9-4bd3-bb9a-a0877ac0db04", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 877.402859] env[63088]: DEBUG oslo_vmware.api [None req-6c4b10e7-44a6-41b3-b172-43d52124d161 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': task-1285065, 'name': PowerOffVM_Task, 'duration_secs': 0.194871} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.403148] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c4b10e7-44a6-41b3-b172-43d52124d161 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: 58d561fe-26aa-4e94-8d55-cc70c361b479] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 877.408512] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-6c4b10e7-44a6-41b3-b172-43d52124d161 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: 58d561fe-26aa-4e94-8d55-cc70c361b479] Reconfiguring VM instance instance-00000045 to detach disk 2001 {{(pid=63088) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 877.409177] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dc2bfbd2-de64-4a80-85c4-6c45744d9455 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.436015] env[63088]: DEBUG oslo_vmware.api [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]5203369c-b1d0-a7dd-9995-995f8f5cf729, 'name': SearchDatastore_Task, 'duration_secs': 0.008928} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.436015] env[63088]: DEBUG oslo_vmware.api [None req-6c4b10e7-44a6-41b3-b172-43d52124d161 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Waiting for the task: (returnval){ [ 877.436015] env[63088]: value = "task-1285068" [ 877.436015] env[63088]: _type = "Task" [ 877.436015] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.436015] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8360996f-85ba-499e-8c12-dddd3ab1cc3c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.449075] env[63088]: DEBUG oslo_vmware.api [None req-6c4b10e7-44a6-41b3-b172-43d52124d161 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': task-1285068, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.450352] env[63088]: DEBUG oslo_vmware.api [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Waiting for the task: (returnval){ [ 877.450352] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]528b4964-65fa-0e13-2a3f-5b99282768f7" [ 877.450352] env[63088]: _type = "Task" [ 877.450352] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.461168] env[63088]: DEBUG oslo_vmware.api [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]528b4964-65fa-0e13-2a3f-5b99282768f7, 'name': SearchDatastore_Task, 'duration_secs': 0.010176} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.461168] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.461168] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] 0fdb21d1-4111-4ff3-bdc0-e2598298a9a4/0fdb21d1-4111-4ff3-bdc0-e2598298a9a4.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 877.461168] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-effe4e31-8cb5-4f4e-8e7b-b65c4e8b9323 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.467296] env[63088]: DEBUG oslo_vmware.api [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Waiting for the task: (returnval){ [ 877.467296] env[63088]: value = "task-1285069" [ 877.467296] env[63088]: _type = "Task" [ 877.467296] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.474907] env[63088]: DEBUG oslo_concurrency.lockutils [None req-2790830b-05cd-47fb-bd8c-69ead0f38b2f tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Lock "7f341d80-450a-47b8-b26d-15f2ce3e378c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.192s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.477774] env[63088]: DEBUG oslo_vmware.api [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285069, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.495123] env[63088]: DEBUG oslo_vmware.api [None req-0dea949d-9593-45db-b8f6-0d07193e67be tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Task: {'id': task-1285067, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.151813} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.495549] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-0dea949d-9593-45db-b8f6-0d07193e67be tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 877.495850] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-0dea949d-9593-45db-b8f6-0d07193e67be tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] [instance: 34ce411f-40c8-446e-b685-cd3ce07663dd] Deleted contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 877.496161] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-0dea949d-9593-45db-b8f6-0d07193e67be tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] [instance: 34ce411f-40c8-446e-b685-cd3ce07663dd] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 877.496448] env[63088]: INFO nova.compute.manager [None req-0dea949d-9593-45db-b8f6-0d07193e67be tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] [instance: 34ce411f-40c8-446e-b685-cd3ce07663dd] Took 1.13 seconds to destroy the instance on the hypervisor. [ 877.496792] env[63088]: DEBUG oslo.service.loopingcall [None req-0dea949d-9593-45db-b8f6-0d07193e67be tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 877.497124] env[63088]: DEBUG nova.compute.manager [-] [instance: 34ce411f-40c8-446e-b685-cd3ce07663dd] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 877.497317] env[63088]: DEBUG nova.network.neutron [-] [instance: 34ce411f-40c8-446e-b685-cd3ce07663dd] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 877.679532] env[63088]: DEBUG nova.compute.manager [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 877.759704] env[63088]: DEBUG oslo_concurrency.lockutils [None req-51a6c6bb-2c09-4c64-956d-a4e00c94e93e tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Releasing lock "refresh_cache-4971b24c-6710-4f50-9846-727dad264b1f" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.760522] env[63088]: DEBUG oslo_concurrency.lockutils [None req-51a6c6bb-2c09-4c64-956d-a4e00c94e93e tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquiring lock "4971b24c-6710-4f50-9846-727dad264b1f" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 877.760697] env[63088]: DEBUG oslo_concurrency.lockutils [None req-51a6c6bb-2c09-4c64-956d-a4e00c94e93e tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquired lock "4971b24c-6710-4f50-9846-727dad264b1f" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.761722] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f13119f-0115-4243-b468-0387b7d66840 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.783964] env[63088]: DEBUG nova.virt.hardware [None req-51a6c6bb-2c09-4c64-956d-a4e00c94e93e tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 877.784284] env[63088]: DEBUG nova.virt.hardware [None req-51a6c6bb-2c09-4c64-956d-a4e00c94e93e tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 877.784489] env[63088]: DEBUG nova.virt.hardware [None req-51a6c6bb-2c09-4c64-956d-a4e00c94e93e tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 877.784719] env[63088]: DEBUG nova.virt.hardware [None req-51a6c6bb-2c09-4c64-956d-a4e00c94e93e tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 877.784912] env[63088]: DEBUG nova.virt.hardware [None req-51a6c6bb-2c09-4c64-956d-a4e00c94e93e tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 877.785114] env[63088]: DEBUG nova.virt.hardware [None req-51a6c6bb-2c09-4c64-956d-a4e00c94e93e tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 877.785384] env[63088]: DEBUG nova.virt.hardware [None req-51a6c6bb-2c09-4c64-956d-a4e00c94e93e tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 877.785598] env[63088]: DEBUG nova.virt.hardware [None req-51a6c6bb-2c09-4c64-956d-a4e00c94e93e tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 877.785810] env[63088]: DEBUG nova.virt.hardware [None req-51a6c6bb-2c09-4c64-956d-a4e00c94e93e tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 877.786012] env[63088]: DEBUG nova.virt.hardware [None req-51a6c6bb-2c09-4c64-956d-a4e00c94e93e tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 877.786239] env[63088]: DEBUG nova.virt.hardware [None req-51a6c6bb-2c09-4c64-956d-a4e00c94e93e tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 877.794494] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-51a6c6bb-2c09-4c64-956d-a4e00c94e93e tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Reconfiguring VM to attach interface {{(pid=63088) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 877.794998] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e3fe070c-7cdb-4631-b66a-0d4874529522 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.814158] env[63088]: DEBUG oslo_vmware.api [None req-51a6c6bb-2c09-4c64-956d-a4e00c94e93e tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Waiting for the task: (returnval){ [ 877.814158] env[63088]: value = "task-1285070" [ 877.814158] env[63088]: _type = "Task" [ 877.814158] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.823250] env[63088]: DEBUG oslo_vmware.api [None req-51a6c6bb-2c09-4c64-956d-a4e00c94e93e tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285070, 'name': ReconfigVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.959480] env[63088]: DEBUG oslo_vmware.api [None req-6c4b10e7-44a6-41b3-b172-43d52124d161 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': task-1285068, 'name': ReconfigVM_Task, 'duration_secs': 0.270475} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.961277] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-6c4b10e7-44a6-41b3-b172-43d52124d161 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: 58d561fe-26aa-4e94-8d55-cc70c361b479] Reconfigured VM instance instance-00000045 to detach disk 2001 {{(pid=63088) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 877.963617] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c4b10e7-44a6-41b3-b172-43d52124d161 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: 58d561fe-26aa-4e94-8d55-cc70c361b479] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 877.963617] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cb996e95-0e4f-454a-a407-ecf4ae99a75c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.970336] env[63088]: DEBUG oslo_vmware.api [None req-6c4b10e7-44a6-41b3-b172-43d52124d161 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Waiting for the task: (returnval){ [ 877.970336] env[63088]: value = "task-1285071" [ 877.970336] env[63088]: _type = "Task" [ 877.970336] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.983367] env[63088]: DEBUG oslo_vmware.api [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285069, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.990043] env[63088]: DEBUG oslo_vmware.api [None req-6c4b10e7-44a6-41b3-b172-43d52124d161 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': task-1285071, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.104707] env[63088]: DEBUG nova.compute.manager [req-188fea46-d4b9-4f3b-92a9-d5e063ff824f req-d54d2ca1-e07d-445e-831e-f2d485556678 service nova] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Received event network-changed-592cc24a-92a9-4bd3-bb9a-a0877ac0db04 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 878.104982] env[63088]: DEBUG nova.compute.manager [req-188fea46-d4b9-4f3b-92a9-d5e063ff824f req-d54d2ca1-e07d-445e-831e-f2d485556678 service nova] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Refreshing instance network info cache due to event network-changed-592cc24a-92a9-4bd3-bb9a-a0877ac0db04. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 878.105772] env[63088]: DEBUG oslo_concurrency.lockutils [req-188fea46-d4b9-4f3b-92a9-d5e063ff824f req-d54d2ca1-e07d-445e-831e-f2d485556678 service nova] Acquiring lock "refresh_cache-4971b24c-6710-4f50-9846-727dad264b1f" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 878.105772] env[63088]: DEBUG oslo_concurrency.lockutils [req-188fea46-d4b9-4f3b-92a9-d5e063ff824f req-d54d2ca1-e07d-445e-831e-f2d485556678 service nova] Acquired lock "refresh_cache-4971b24c-6710-4f50-9846-727dad264b1f" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.105772] env[63088]: DEBUG nova.network.neutron [req-188fea46-d4b9-4f3b-92a9-d5e063ff824f req-d54d2ca1-e07d-445e-831e-f2d485556678 service nova] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Refreshing network info cache for port 592cc24a-92a9-4bd3-bb9a-a0877ac0db04 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 878.218514] env[63088]: DEBUG oslo_concurrency.lockutils [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 878.297634] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dab20eae-16f1-4395-82c2-6fcf6d484cc6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.311195] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48473636-c1b7-4280-a095-5b3c8a1a80ae {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.333603] env[63088]: DEBUG oslo_vmware.api [None req-51a6c6bb-2c09-4c64-956d-a4e00c94e93e tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285070, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.380428] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecfb6aef-2429-44fb-979d-fc38d2d9df34 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.392932] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f454ad24-a1cc-4572-8e83-5914fdadb365 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.412515] env[63088]: DEBUG nova.compute.provider_tree [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 878.487305] env[63088]: DEBUG oslo_vmware.api [None req-6c4b10e7-44a6-41b3-b172-43d52124d161 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': task-1285071, 'name': PowerOnVM_Task, 'duration_secs': 0.494002} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.490769] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c4b10e7-44a6-41b3-b172-43d52124d161 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: 58d561fe-26aa-4e94-8d55-cc70c361b479] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 878.491092] env[63088]: DEBUG nova.compute.manager [None req-6c4b10e7-44a6-41b3-b172-43d52124d161 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: 58d561fe-26aa-4e94-8d55-cc70c361b479] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 878.491989] env[63088]: DEBUG oslo_vmware.api [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285069, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.51669} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.492887] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce3cdb05-6582-4a16-a9d2-0d3c55479928 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.495982] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] 0fdb21d1-4111-4ff3-bdc0-e2598298a9a4/0fdb21d1-4111-4ff3-bdc0-e2598298a9a4.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 878.496134] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 0fdb21d1-4111-4ff3-bdc0-e2598298a9a4] Extending root virtual disk to 1048576 {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 878.496399] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-237e5a87-98a7-411b-965e-94d6de3548c4 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.505304] env[63088]: DEBUG oslo_vmware.api [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Waiting for the task: (returnval){ [ 878.505304] env[63088]: value = "task-1285072" [ 878.505304] env[63088]: _type = "Task" [ 878.505304] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.513234] env[63088]: DEBUG oslo_vmware.api [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285072, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.719194] env[63088]: DEBUG nova.network.neutron [-] [instance: 34ce411f-40c8-446e-b685-cd3ce07663dd] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 878.727027] env[63088]: DEBUG oslo_concurrency.lockutils [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Acquiring lock "bebc3318-24bb-4a37-8b23-66a12a7f7fd2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 878.727027] env[63088]: DEBUG oslo_concurrency.lockutils [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Lock "bebc3318-24bb-4a37-8b23-66a12a7f7fd2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 878.776031] env[63088]: DEBUG oslo_concurrency.lockutils [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Acquiring lock "45513c6b-c000-4ee1-8893-4e084ffc22c5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 878.776374] env[63088]: DEBUG oslo_concurrency.lockutils [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Lock "45513c6b-c000-4ee1-8893-4e084ffc22c5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.004s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 878.830327] env[63088]: DEBUG oslo_vmware.api [None req-51a6c6bb-2c09-4c64-956d-a4e00c94e93e tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285070, 'name': ReconfigVM_Task, 'duration_secs': 0.760776} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.830937] env[63088]: DEBUG oslo_concurrency.lockutils [None req-51a6c6bb-2c09-4c64-956d-a4e00c94e93e tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Releasing lock "4971b24c-6710-4f50-9846-727dad264b1f" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 878.831218] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-51a6c6bb-2c09-4c64-956d-a4e00c94e93e tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Reconfigured VM to attach interface {{(pid=63088) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 878.915548] env[63088]: DEBUG nova.scheduler.client.report [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 879.021194] env[63088]: DEBUG oslo_vmware.api [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285072, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.133984} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.021519] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 0fdb21d1-4111-4ff3-bdc0-e2598298a9a4] Extended root virtual disk {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 879.022410] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ab28073-f952-46c1-be7e-98df8496d945 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.047254] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 0fdb21d1-4111-4ff3-bdc0-e2598298a9a4] Reconfiguring VM instance instance-00000048 to attach disk [datastore1] 0fdb21d1-4111-4ff3-bdc0-e2598298a9a4/0fdb21d1-4111-4ff3-bdc0-e2598298a9a4.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 879.048965] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-24565c48-2c1c-48cc-9d73-37bbd15bd785 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 9d5f1cde-e787-4a2c-ac49-83ec135ff51c] Volume attach. Driver type: vmdk {{(pid=63088) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 879.049193] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-24565c48-2c1c-48cc-9d73-37bbd15bd785 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 9d5f1cde-e787-4a2c-ac49-83ec135ff51c] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-275890', 'volume_id': 'b674aa10-140a-4cb0-8ca0-7fa9fddf7a6d', 'name': 'volume-b674aa10-140a-4cb0-8ca0-7fa9fddf7a6d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9d5f1cde-e787-4a2c-ac49-83ec135ff51c', 'attached_at': '', 'detached_at': '', 'volume_id': 'b674aa10-140a-4cb0-8ca0-7fa9fddf7a6d', 'serial': 'b674aa10-140a-4cb0-8ca0-7fa9fddf7a6d'} {{(pid=63088) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 879.049486] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4f72297d-c47c-49ff-b628-ce44c4065cd5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.065126] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75a21f61-f90d-41c4-bad8-69ebd0904052 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.087433] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fd884d0-a837-45c5-b53a-b814588b7aab {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.090774] env[63088]: DEBUG oslo_vmware.api [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Waiting for the task: (returnval){ [ 879.090774] env[63088]: value = "task-1285073" [ 879.090774] env[63088]: _type = "Task" [ 879.090774] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.117314] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-24565c48-2c1c-48cc-9d73-37bbd15bd785 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 9d5f1cde-e787-4a2c-ac49-83ec135ff51c] Reconfiguring VM instance instance-00000046 to attach disk [datastore1] volume-b674aa10-140a-4cb0-8ca0-7fa9fddf7a6d/volume-b674aa10-140a-4cb0-8ca0-7fa9fddf7a6d.vmdk or device None with type thin {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 879.118130] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0f622f73-3926-4fd5-a231-59d96c1c47ab {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.134021] env[63088]: DEBUG nova.network.neutron [req-188fea46-d4b9-4f3b-92a9-d5e063ff824f req-d54d2ca1-e07d-445e-831e-f2d485556678 service nova] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Updated VIF entry in instance network info cache for port 592cc24a-92a9-4bd3-bb9a-a0877ac0db04. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 879.134021] env[63088]: DEBUG nova.network.neutron [req-188fea46-d4b9-4f3b-92a9-d5e063ff824f req-d54d2ca1-e07d-445e-831e-f2d485556678 service nova] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Updating instance_info_cache with network_info: [{"id": "d5a10951-ae2e-48f5-b0ca-b1f144e5fe3a", "address": "fa:16:3e:95:96:0c", "network": {"id": "1174a284-d3ca-4f9e-aa81-13ee9a693e55", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1994276040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.166", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa3d24a1a6c0430985fd80365d986ee1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd5a10951-ae", "ovs_interfaceid": "d5a10951-ae2e-48f5-b0ca-b1f144e5fe3a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "81932288-bc32-47c4-a3b2-02597cb6bf81", "address": "fa:16:3e:17:d1:1a", "network": {"id": "1174a284-d3ca-4f9e-aa81-13ee9a693e55", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1994276040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa3d24a1a6c0430985fd80365d986ee1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap81932288-bc", "ovs_interfaceid": "81932288-bc32-47c4-a3b2-02597cb6bf81", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "592cc24a-92a9-4bd3-bb9a-a0877ac0db04", "address": "fa:16:3e:fe:d7:1a", "network": {"id": "1174a284-d3ca-4f9e-aa81-13ee9a693e55", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1994276040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa3d24a1a6c0430985fd80365d986ee1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap592cc24a-92", "ovs_interfaceid": "592cc24a-92a9-4bd3-bb9a-a0877ac0db04", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 879.138010] env[63088]: DEBUG oslo_vmware.api [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285073, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.142191] env[63088]: DEBUG oslo_vmware.api [None req-24565c48-2c1c-48cc-9d73-37bbd15bd785 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 879.142191] env[63088]: value = "task-1285074" [ 879.142191] env[63088]: _type = "Task" [ 879.142191] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.150435] env[63088]: DEBUG oslo_vmware.api [None req-24565c48-2c1c-48cc-9d73-37bbd15bd785 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285074, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.225846] env[63088]: INFO nova.compute.manager [-] [instance: 34ce411f-40c8-446e-b685-cd3ce07663dd] Took 1.73 seconds to deallocate network for instance. [ 879.228012] env[63088]: DEBUG nova.compute.manager [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: bebc3318-24bb-4a37-8b23-66a12a7f7fd2] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 879.281101] env[63088]: DEBUG nova.compute.manager [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 45513c6b-c000-4ee1-8893-4e084ffc22c5] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 879.338299] env[63088]: DEBUG oslo_concurrency.lockutils [None req-51a6c6bb-2c09-4c64-956d-a4e00c94e93e tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Lock "interface-4971b24c-6710-4f50-9846-727dad264b1f-592cc24a-92a9-4bd3-bb9a-a0877ac0db04" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.740s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.420729] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.483s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.421284] env[63088]: DEBUG nova.compute.manager [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: 667fd5e9-5fe4-41e1-9d8a-896c6e5c6286] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 879.424134] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.717s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.427816] env[63088]: INFO nova.compute.claims [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: b50db882-598c-488f-b935-34f55f655642] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 879.601658] env[63088]: DEBUG oslo_vmware.api [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285073, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.638902] env[63088]: DEBUG oslo_concurrency.lockutils [req-188fea46-d4b9-4f3b-92a9-d5e063ff824f req-d54d2ca1-e07d-445e-831e-f2d485556678 service nova] Releasing lock "refresh_cache-4971b24c-6710-4f50-9846-727dad264b1f" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 879.654448] env[63088]: DEBUG oslo_vmware.api [None req-24565c48-2c1c-48cc-9d73-37bbd15bd785 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285074, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.737216] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0dea949d-9593-45db-b8f6-0d07193e67be tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.759128] env[63088]: DEBUG oslo_concurrency.lockutils [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.800353] env[63088]: DEBUG oslo_concurrency.lockutils [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.820325] env[63088]: DEBUG oslo_concurrency.lockutils [None req-e757b386-7dab-43e3-b44f-3eab829d5a35 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Acquiring lock "58d561fe-26aa-4e94-8d55-cc70c361b479" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.820848] env[63088]: DEBUG oslo_concurrency.lockutils [None req-e757b386-7dab-43e3-b44f-3eab829d5a35 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Lock "58d561fe-26aa-4e94-8d55-cc70c361b479" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.821258] env[63088]: DEBUG oslo_concurrency.lockutils [None req-e757b386-7dab-43e3-b44f-3eab829d5a35 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Acquiring lock "58d561fe-26aa-4e94-8d55-cc70c361b479-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.821585] env[63088]: DEBUG oslo_concurrency.lockutils [None req-e757b386-7dab-43e3-b44f-3eab829d5a35 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Lock "58d561fe-26aa-4e94-8d55-cc70c361b479-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.821897] env[63088]: DEBUG oslo_concurrency.lockutils [None req-e757b386-7dab-43e3-b44f-3eab829d5a35 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Lock "58d561fe-26aa-4e94-8d55-cc70c361b479-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.824522] env[63088]: INFO nova.compute.manager [None req-e757b386-7dab-43e3-b44f-3eab829d5a35 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: 58d561fe-26aa-4e94-8d55-cc70c361b479] Terminating instance [ 879.829137] env[63088]: DEBUG nova.compute.manager [None req-e757b386-7dab-43e3-b44f-3eab829d5a35 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: 58d561fe-26aa-4e94-8d55-cc70c361b479] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 879.829137] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-e757b386-7dab-43e3-b44f-3eab829d5a35 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: 58d561fe-26aa-4e94-8d55-cc70c361b479] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 879.829137] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc966160-391f-4d47-9735-179234fc2176 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.838450] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-e757b386-7dab-43e3-b44f-3eab829d5a35 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: 58d561fe-26aa-4e94-8d55-cc70c361b479] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 879.838911] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-15ac174f-1dd6-4ba6-9fee-4e55b5e892e8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.846759] env[63088]: DEBUG oslo_vmware.api [None req-e757b386-7dab-43e3-b44f-3eab829d5a35 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Waiting for the task: (returnval){ [ 879.846759] env[63088]: value = "task-1285075" [ 879.846759] env[63088]: _type = "Task" [ 879.846759] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.855205] env[63088]: DEBUG oslo_vmware.api [None req-e757b386-7dab-43e3-b44f-3eab829d5a35 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': task-1285075, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.930113] env[63088]: DEBUG nova.compute.utils [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 879.934599] env[63088]: DEBUG nova.compute.manager [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: 667fd5e9-5fe4-41e1-9d8a-896c6e5c6286] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 879.934599] env[63088]: DEBUG nova.network.neutron [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: 667fd5e9-5fe4-41e1-9d8a-896c6e5c6286] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 879.987249] env[63088]: DEBUG nova.policy [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '065ecdba92ea4c5a9f6f493eefd6fa51', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a421ece306c64913b7cf9b55709a49c5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 880.101803] env[63088]: DEBUG oslo_vmware.api [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285073, 'name': ReconfigVM_Task, 'duration_secs': 1.002067} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.102152] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 0fdb21d1-4111-4ff3-bdc0-e2598298a9a4] Reconfigured VM instance instance-00000048 to attach disk [datastore1] 0fdb21d1-4111-4ff3-bdc0-e2598298a9a4/0fdb21d1-4111-4ff3-bdc0-e2598298a9a4.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 880.102800] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5b736a82-e15a-40cf-b37a-08fa0591161b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.108873] env[63088]: DEBUG oslo_vmware.api [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Waiting for the task: (returnval){ [ 880.108873] env[63088]: value = "task-1285076" [ 880.108873] env[63088]: _type = "Task" [ 880.108873] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.117246] env[63088]: DEBUG oslo_vmware.api [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285076, 'name': Rename_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.153526] env[63088]: DEBUG oslo_vmware.api [None req-24565c48-2c1c-48cc-9d73-37bbd15bd785 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285074, 'name': ReconfigVM_Task, 'duration_secs': 0.93488} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.153853] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-24565c48-2c1c-48cc-9d73-37bbd15bd785 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 9d5f1cde-e787-4a2c-ac49-83ec135ff51c] Reconfigured VM instance instance-00000046 to attach disk [datastore1] volume-b674aa10-140a-4cb0-8ca0-7fa9fddf7a6d/volume-b674aa10-140a-4cb0-8ca0-7fa9fddf7a6d.vmdk or device None with type thin {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 880.158700] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e4d3134f-9776-40b3-8b47-64f7da22ae25 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.174490] env[63088]: DEBUG oslo_vmware.api [None req-24565c48-2c1c-48cc-9d73-37bbd15bd785 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 880.174490] env[63088]: value = "task-1285077" [ 880.174490] env[63088]: _type = "Task" [ 880.174490] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.183782] env[63088]: DEBUG oslo_vmware.api [None req-24565c48-2c1c-48cc-9d73-37bbd15bd785 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285077, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.230448] env[63088]: DEBUG nova.compute.manager [req-af90508e-476c-4df7-8556-230538f4b73b req-301e292f-e12e-450d-a7fb-444fa663c098 service nova] [instance: 34ce411f-40c8-446e-b685-cd3ce07663dd] Received event network-vif-deleted-44d58f1e-f90f-4aea-b8fd-5f9c7ff195cb {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 880.359295] env[63088]: DEBUG oslo_vmware.api [None req-e757b386-7dab-43e3-b44f-3eab829d5a35 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': task-1285075, 'name': PowerOffVM_Task, 'duration_secs': 0.199611} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.359504] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-e757b386-7dab-43e3-b44f-3eab829d5a35 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: 58d561fe-26aa-4e94-8d55-cc70c361b479] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 880.359672] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-e757b386-7dab-43e3-b44f-3eab829d5a35 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: 58d561fe-26aa-4e94-8d55-cc70c361b479] Unregistering the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 880.360410] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d897779e-0cbe-4118-acf0-013200349381 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.383617] env[63088]: DEBUG nova.network.neutron [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: 667fd5e9-5fe4-41e1-9d8a-896c6e5c6286] Successfully created port: 602aeee4-c8d8-41c6-b145-dae2eadfc9a7 {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 880.423633] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-e757b386-7dab-43e3-b44f-3eab829d5a35 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: 58d561fe-26aa-4e94-8d55-cc70c361b479] Unregistered the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 880.423633] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-e757b386-7dab-43e3-b44f-3eab829d5a35 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: 58d561fe-26aa-4e94-8d55-cc70c361b479] Deleting contents of the VM from datastore datastore2 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 880.423633] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-e757b386-7dab-43e3-b44f-3eab829d5a35 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Deleting the datastore file [datastore2] 58d561fe-26aa-4e94-8d55-cc70c361b479 {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 880.423633] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9c5989be-e701-4615-bbbe-1b73983a8707 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.430917] env[63088]: DEBUG oslo_vmware.api [None req-e757b386-7dab-43e3-b44f-3eab829d5a35 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Waiting for the task: (returnval){ [ 880.430917] env[63088]: value = "task-1285079" [ 880.430917] env[63088]: _type = "Task" [ 880.430917] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.435718] env[63088]: DEBUG nova.compute.manager [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: 667fd5e9-5fe4-41e1-9d8a-896c6e5c6286] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 880.450846] env[63088]: DEBUG oslo_vmware.api [None req-e757b386-7dab-43e3-b44f-3eab829d5a35 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': task-1285079, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.620645] env[63088]: DEBUG oslo_vmware.api [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285076, 'name': Rename_Task, 'duration_secs': 0.129864} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.622951] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 0fdb21d1-4111-4ff3-bdc0-e2598298a9a4] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 880.623870] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-26cdc7b1-2999-4677-9f86-9648fb266fc1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.629110] env[63088]: DEBUG oslo_vmware.api [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Waiting for the task: (returnval){ [ 880.629110] env[63088]: value = "task-1285080" [ 880.629110] env[63088]: _type = "Task" [ 880.629110] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.639462] env[63088]: DEBUG oslo_vmware.api [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285080, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.688049] env[63088]: DEBUG oslo_vmware.api [None req-24565c48-2c1c-48cc-9d73-37bbd15bd785 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285077, 'name': ReconfigVM_Task, 'duration_secs': 0.150133} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.688371] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-24565c48-2c1c-48cc-9d73-37bbd15bd785 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 9d5f1cde-e787-4a2c-ac49-83ec135ff51c] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-275890', 'volume_id': 'b674aa10-140a-4cb0-8ca0-7fa9fddf7a6d', 'name': 'volume-b674aa10-140a-4cb0-8ca0-7fa9fddf7a6d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9d5f1cde-e787-4a2c-ac49-83ec135ff51c', 'attached_at': '', 'detached_at': '', 'volume_id': 'b674aa10-140a-4cb0-8ca0-7fa9fddf7a6d', 'serial': 'b674aa10-140a-4cb0-8ca0-7fa9fddf7a6d'} {{(pid=63088) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 880.735527] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68f89361-625d-4f87-9dcc-847a96b7b247 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.747074] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edbd46bb-b436-47f2-9be2-dee4f4d2b843 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.776508] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f87f0be-b51e-4e8c-8287-cfd3ef3d7a9d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.784286] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a583bd08-e62a-4e6c-beaf-88f3df97bf22 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.797120] env[63088]: DEBUG nova.compute.provider_tree [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 880.946125] env[63088]: DEBUG oslo_vmware.api [None req-e757b386-7dab-43e3-b44f-3eab829d5a35 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': task-1285079, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.152203} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.946603] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-e757b386-7dab-43e3-b44f-3eab829d5a35 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 880.947072] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-e757b386-7dab-43e3-b44f-3eab829d5a35 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: 58d561fe-26aa-4e94-8d55-cc70c361b479] Deleted contents of the VM from datastore datastore2 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 880.947072] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-e757b386-7dab-43e3-b44f-3eab829d5a35 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: 58d561fe-26aa-4e94-8d55-cc70c361b479] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 880.947239] env[63088]: INFO nova.compute.manager [None req-e757b386-7dab-43e3-b44f-3eab829d5a35 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: 58d561fe-26aa-4e94-8d55-cc70c361b479] Took 1.12 seconds to destroy the instance on the hypervisor. [ 880.947406] env[63088]: DEBUG oslo.service.loopingcall [None req-e757b386-7dab-43e3-b44f-3eab829d5a35 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 880.947685] env[63088]: DEBUG nova.compute.manager [-] [instance: 58d561fe-26aa-4e94-8d55-cc70c361b479] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 880.947685] env[63088]: DEBUG nova.network.neutron [-] [instance: 58d561fe-26aa-4e94-8d55-cc70c361b479] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 881.085182] env[63088]: DEBUG oslo_concurrency.lockutils [None req-1ece652d-e3a1-4b82-be99-c1747ddf14f1 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquiring lock "interface-4971b24c-6710-4f50-9846-727dad264b1f-81932288-bc32-47c4-a3b2-02597cb6bf81" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.085300] env[63088]: DEBUG oslo_concurrency.lockutils [None req-1ece652d-e3a1-4b82-be99-c1747ddf14f1 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Lock "interface-4971b24c-6710-4f50-9846-727dad264b1f-81932288-bc32-47c4-a3b2-02597cb6bf81" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.138762] env[63088]: DEBUG oslo_vmware.api [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285080, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.300515] env[63088]: DEBUG nova.scheduler.client.report [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 881.449684] env[63088]: DEBUG nova.compute.manager [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: 667fd5e9-5fe4-41e1-9d8a-896c6e5c6286] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 881.476339] env[63088]: DEBUG nova.virt.hardware [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 881.476590] env[63088]: DEBUG nova.virt.hardware [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 881.476748] env[63088]: DEBUG nova.virt.hardware [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 881.477097] env[63088]: DEBUG nova.virt.hardware [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 881.477263] env[63088]: DEBUG nova.virt.hardware [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 881.477417] env[63088]: DEBUG nova.virt.hardware [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 881.477625] env[63088]: DEBUG nova.virt.hardware [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 881.477784] env[63088]: DEBUG nova.virt.hardware [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 881.477961] env[63088]: DEBUG nova.virt.hardware [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 881.478186] env[63088]: DEBUG nova.virt.hardware [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 881.478467] env[63088]: DEBUG nova.virt.hardware [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 881.479358] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec0f8e05-2595-4487-a9c3-a61e47716ef7 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.487734] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-746e4708-82a5-4f9e-89f4-d0451cf5e285 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.591405] env[63088]: DEBUG oslo_concurrency.lockutils [None req-1ece652d-e3a1-4b82-be99-c1747ddf14f1 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquiring lock "4971b24c-6710-4f50-9846-727dad264b1f" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 881.591609] env[63088]: DEBUG oslo_concurrency.lockutils [None req-1ece652d-e3a1-4b82-be99-c1747ddf14f1 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquired lock "4971b24c-6710-4f50-9846-727dad264b1f" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 881.592596] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6268fbcd-01e1-400e-887b-ba8cf163d702 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.611642] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74c9bd35-3a15-4313-8fa7-7fd0f4af0503 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.639821] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-1ece652d-e3a1-4b82-be99-c1747ddf14f1 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Reconfiguring VM to detach interface {{(pid=63088) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1974}} [ 881.643020] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7d0bf67e-7743-4751-9d4e-2f2de9dc1b40 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.663265] env[63088]: DEBUG oslo_vmware.api [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285080, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.664683] env[63088]: DEBUG oslo_vmware.api [None req-1ece652d-e3a1-4b82-be99-c1747ddf14f1 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Waiting for the task: (returnval){ [ 881.664683] env[63088]: value = "task-1285081" [ 881.664683] env[63088]: _type = "Task" [ 881.664683] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.667871] env[63088]: DEBUG nova.network.neutron [-] [instance: 58d561fe-26aa-4e94-8d55-cc70c361b479] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 881.674629] env[63088]: DEBUG oslo_vmware.api [None req-1ece652d-e3a1-4b82-be99-c1747ddf14f1 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285081, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.734129] env[63088]: DEBUG nova.objects.instance [None req-24565c48-2c1c-48cc-9d73-37bbd15bd785 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lazy-loading 'flavor' on Instance uuid 9d5f1cde-e787-4a2c-ac49-83ec135ff51c {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 881.806725] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.382s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.807319] env[63088]: DEBUG nova.compute.manager [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: b50db882-598c-488f-b935-34f55f655642] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 881.810048] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.073s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.811979] env[63088]: INFO nova.compute.claims [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: 2964db35-7357-40a7-b4e6-7e2595549f5b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 881.924515] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c799bf18-c9a4-4b67-989b-6dda4cd4483c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquiring lock "9d5f1cde-e787-4a2c-ac49-83ec135ff51c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.067217] env[63088]: DEBUG nova.network.neutron [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: 667fd5e9-5fe4-41e1-9d8a-896c6e5c6286] Successfully updated port: 602aeee4-c8d8-41c6-b145-dae2eadfc9a7 {{(pid=63088) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 882.145171] env[63088]: DEBUG oslo_vmware.api [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285080, 'name': PowerOnVM_Task, 'duration_secs': 1.261081} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.145762] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 0fdb21d1-4111-4ff3-bdc0-e2598298a9a4] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 882.145970] env[63088]: INFO nova.compute.manager [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 0fdb21d1-4111-4ff3-bdc0-e2598298a9a4] Took 9.84 seconds to spawn the instance on the hypervisor. [ 882.146174] env[63088]: DEBUG nova.compute.manager [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 0fdb21d1-4111-4ff3-bdc0-e2598298a9a4] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 882.147015] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40cc1989-fb57-43b8-a620-437b9dd2c92f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.170599] env[63088]: INFO nova.compute.manager [-] [instance: 58d561fe-26aa-4e94-8d55-cc70c361b479] Took 1.22 seconds to deallocate network for instance. [ 882.175334] env[63088]: DEBUG oslo_vmware.api [None req-1ece652d-e3a1-4b82-be99-c1747ddf14f1 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285081, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.239171] env[63088]: DEBUG oslo_concurrency.lockutils [None req-24565c48-2c1c-48cc-9d73-37bbd15bd785 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "9d5f1cde-e787-4a2c-ac49-83ec135ff51c" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 9.838s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.240181] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c799bf18-c9a4-4b67-989b-6dda4cd4483c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "9d5f1cde-e787-4a2c-ac49-83ec135ff51c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.316s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.240414] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c799bf18-c9a4-4b67-989b-6dda4cd4483c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquiring lock "9d5f1cde-e787-4a2c-ac49-83ec135ff51c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.240632] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c799bf18-c9a4-4b67-989b-6dda4cd4483c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "9d5f1cde-e787-4a2c-ac49-83ec135ff51c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.240809] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c799bf18-c9a4-4b67-989b-6dda4cd4483c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "9d5f1cde-e787-4a2c-ac49-83ec135ff51c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.242489] env[63088]: INFO nova.compute.manager [None req-c799bf18-c9a4-4b67-989b-6dda4cd4483c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 9d5f1cde-e787-4a2c-ac49-83ec135ff51c] Terminating instance [ 882.244823] env[63088]: DEBUG nova.compute.manager [None req-c799bf18-c9a4-4b67-989b-6dda4cd4483c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 9d5f1cde-e787-4a2c-ac49-83ec135ff51c] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 882.245059] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-c799bf18-c9a4-4b67-989b-6dda4cd4483c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 9d5f1cde-e787-4a2c-ac49-83ec135ff51c] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 882.245305] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-73ad74ec-eaaf-4de3-adae-dcf3c682bac3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.261724] env[63088]: DEBUG nova.compute.manager [req-0e5dd796-c09c-42e5-957a-fd7d271d0e61 req-4c78697f-0b8d-4e42-8e6a-dbc49a97328a service nova] [instance: 58d561fe-26aa-4e94-8d55-cc70c361b479] Received event network-vif-deleted-c6477c64-3129-4546-84b1-40bc79bcaa57 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 882.261933] env[63088]: DEBUG nova.compute.manager [req-0e5dd796-c09c-42e5-957a-fd7d271d0e61 req-4c78697f-0b8d-4e42-8e6a-dbc49a97328a service nova] [instance: 667fd5e9-5fe4-41e1-9d8a-896c6e5c6286] Received event network-vif-plugged-602aeee4-c8d8-41c6-b145-dae2eadfc9a7 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 882.262146] env[63088]: DEBUG oslo_concurrency.lockutils [req-0e5dd796-c09c-42e5-957a-fd7d271d0e61 req-4c78697f-0b8d-4e42-8e6a-dbc49a97328a service nova] Acquiring lock "667fd5e9-5fe4-41e1-9d8a-896c6e5c6286-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.262722] env[63088]: DEBUG oslo_concurrency.lockutils [req-0e5dd796-c09c-42e5-957a-fd7d271d0e61 req-4c78697f-0b8d-4e42-8e6a-dbc49a97328a service nova] Lock "667fd5e9-5fe4-41e1-9d8a-896c6e5c6286-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.262904] env[63088]: DEBUG oslo_concurrency.lockutils [req-0e5dd796-c09c-42e5-957a-fd7d271d0e61 req-4c78697f-0b8d-4e42-8e6a-dbc49a97328a service nova] Lock "667fd5e9-5fe4-41e1-9d8a-896c6e5c6286-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.263104] env[63088]: DEBUG nova.compute.manager [req-0e5dd796-c09c-42e5-957a-fd7d271d0e61 req-4c78697f-0b8d-4e42-8e6a-dbc49a97328a service nova] [instance: 667fd5e9-5fe4-41e1-9d8a-896c6e5c6286] No waiting events found dispatching network-vif-plugged-602aeee4-c8d8-41c6-b145-dae2eadfc9a7 {{(pid=63088) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 882.263280] env[63088]: WARNING nova.compute.manager [req-0e5dd796-c09c-42e5-957a-fd7d271d0e61 req-4c78697f-0b8d-4e42-8e6a-dbc49a97328a service nova] [instance: 667fd5e9-5fe4-41e1-9d8a-896c6e5c6286] Received unexpected event network-vif-plugged-602aeee4-c8d8-41c6-b145-dae2eadfc9a7 for instance with vm_state building and task_state spawning. [ 882.263446] env[63088]: DEBUG nova.compute.manager [req-0e5dd796-c09c-42e5-957a-fd7d271d0e61 req-4c78697f-0b8d-4e42-8e6a-dbc49a97328a service nova] [instance: 667fd5e9-5fe4-41e1-9d8a-896c6e5c6286] Received event network-changed-602aeee4-c8d8-41c6-b145-dae2eadfc9a7 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 882.263602] env[63088]: DEBUG nova.compute.manager [req-0e5dd796-c09c-42e5-957a-fd7d271d0e61 req-4c78697f-0b8d-4e42-8e6a-dbc49a97328a service nova] [instance: 667fd5e9-5fe4-41e1-9d8a-896c6e5c6286] Refreshing instance network info cache due to event network-changed-602aeee4-c8d8-41c6-b145-dae2eadfc9a7. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 882.263797] env[63088]: DEBUG oslo_concurrency.lockutils [req-0e5dd796-c09c-42e5-957a-fd7d271d0e61 req-4c78697f-0b8d-4e42-8e6a-dbc49a97328a service nova] Acquiring lock "refresh_cache-667fd5e9-5fe4-41e1-9d8a-896c6e5c6286" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 882.263935] env[63088]: DEBUG oslo_concurrency.lockutils [req-0e5dd796-c09c-42e5-957a-fd7d271d0e61 req-4c78697f-0b8d-4e42-8e6a-dbc49a97328a service nova] Acquired lock "refresh_cache-667fd5e9-5fe4-41e1-9d8a-896c6e5c6286" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 882.264110] env[63088]: DEBUG nova.network.neutron [req-0e5dd796-c09c-42e5-957a-fd7d271d0e61 req-4c78697f-0b8d-4e42-8e6a-dbc49a97328a service nova] [instance: 667fd5e9-5fe4-41e1-9d8a-896c6e5c6286] Refreshing network info cache for port 602aeee4-c8d8-41c6-b145-dae2eadfc9a7 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 882.266641] env[63088]: DEBUG oslo_vmware.api [None req-c799bf18-c9a4-4b67-989b-6dda4cd4483c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 882.266641] env[63088]: value = "task-1285082" [ 882.266641] env[63088]: _type = "Task" [ 882.266641] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.276856] env[63088]: DEBUG oslo_vmware.api [None req-c799bf18-c9a4-4b67-989b-6dda4cd4483c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285082, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.316151] env[63088]: DEBUG nova.compute.utils [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 882.319467] env[63088]: DEBUG nova.compute.manager [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: b50db882-598c-488f-b935-34f55f655642] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 882.319633] env[63088]: DEBUG nova.network.neutron [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: b50db882-598c-488f-b935-34f55f655642] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 882.392502] env[63088]: DEBUG nova.policy [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '065ecdba92ea4c5a9f6f493eefd6fa51', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a421ece306c64913b7cf9b55709a49c5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 882.573282] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Acquiring lock "refresh_cache-667fd5e9-5fe4-41e1-9d8a-896c6e5c6286" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 882.675099] env[63088]: INFO nova.compute.manager [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 0fdb21d1-4111-4ff3-bdc0-e2598298a9a4] Took 33.72 seconds to build instance. [ 882.686172] env[63088]: DEBUG oslo_concurrency.lockutils [None req-e757b386-7dab-43e3-b44f-3eab829d5a35 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.686172] env[63088]: DEBUG oslo_vmware.api [None req-1ece652d-e3a1-4b82-be99-c1747ddf14f1 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285081, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.700932] env[63088]: DEBUG nova.network.neutron [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: b50db882-598c-488f-b935-34f55f655642] Successfully created port: 585051d5-6974-4abd-864e-38967371f048 {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 882.777647] env[63088]: DEBUG oslo_vmware.api [None req-c799bf18-c9a4-4b67-989b-6dda4cd4483c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285082, 'name': PowerOffVM_Task, 'duration_secs': 0.250397} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.777965] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-c799bf18-c9a4-4b67-989b-6dda4cd4483c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 9d5f1cde-e787-4a2c-ac49-83ec135ff51c] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 882.778184] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-c799bf18-c9a4-4b67-989b-6dda4cd4483c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 9d5f1cde-e787-4a2c-ac49-83ec135ff51c] Volume detach. Driver type: vmdk {{(pid=63088) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 882.778390] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-c799bf18-c9a4-4b67-989b-6dda4cd4483c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 9d5f1cde-e787-4a2c-ac49-83ec135ff51c] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-275890', 'volume_id': 'b674aa10-140a-4cb0-8ca0-7fa9fddf7a6d', 'name': 'volume-b674aa10-140a-4cb0-8ca0-7fa9fddf7a6d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9d5f1cde-e787-4a2c-ac49-83ec135ff51c', 'attached_at': '', 'detached_at': '', 'volume_id': 'b674aa10-140a-4cb0-8ca0-7fa9fddf7a6d', 'serial': 'b674aa10-140a-4cb0-8ca0-7fa9fddf7a6d'} {{(pid=63088) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 882.779269] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a12a35a3-31e3-4949-ab68-70367eeac7c3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.804583] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51b4b55d-4d30-4f87-946e-a23e9ca30327 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.814040] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d8b8bf3-0633-43aa-876f-7a9e1eaf2468 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.825243] env[63088]: DEBUG nova.network.neutron [req-0e5dd796-c09c-42e5-957a-fd7d271d0e61 req-4c78697f-0b8d-4e42-8e6a-dbc49a97328a service nova] [instance: 667fd5e9-5fe4-41e1-9d8a-896c6e5c6286] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 882.843607] env[63088]: DEBUG nova.compute.manager [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: b50db882-598c-488f-b935-34f55f655642] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 882.853432] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7912ddf-7e3c-4e6d-9327-20bcfb66d636 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.868144] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-c799bf18-c9a4-4b67-989b-6dda4cd4483c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] The volume has not been displaced from its original location: [datastore1] volume-b674aa10-140a-4cb0-8ca0-7fa9fddf7a6d/volume-b674aa10-140a-4cb0-8ca0-7fa9fddf7a6d.vmdk. No consolidation needed. {{(pid=63088) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 882.873616] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-c799bf18-c9a4-4b67-989b-6dda4cd4483c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 9d5f1cde-e787-4a2c-ac49-83ec135ff51c] Reconfiguring VM instance instance-00000046 to detach disk 2001 {{(pid=63088) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 882.876510] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f18b545d-e3eb-42b4-ac50-86af88f5b5cc {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.897782] env[63088]: DEBUG oslo_vmware.api [None req-c799bf18-c9a4-4b67-989b-6dda4cd4483c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 882.897782] env[63088]: value = "task-1285083" [ 882.897782] env[63088]: _type = "Task" [ 882.897782] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.909949] env[63088]: DEBUG oslo_vmware.api [None req-c799bf18-c9a4-4b67-989b-6dda4cd4483c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285083, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.998825] env[63088]: DEBUG nova.network.neutron [req-0e5dd796-c09c-42e5-957a-fd7d271d0e61 req-4c78697f-0b8d-4e42-8e6a-dbc49a97328a service nova] [instance: 667fd5e9-5fe4-41e1-9d8a-896c6e5c6286] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 883.101031] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f1666685-2914-415b-a4ae-382918967cb8 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquiring lock "0fdb21d1-4111-4ff3-bdc0-e2598298a9a4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 883.162560] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-219edecd-6a11-4c6a-9661-3798371bc87c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.175937] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d19a5e3-f92a-4d96-b26f-588ea8b00b1e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.185355] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b6495512-26ff-4928-9452-f14f8f6f8171 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lock "0fdb21d1-4111-4ff3-bdc0-e2598298a9a4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.238s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 883.185355] env[63088]: DEBUG oslo_vmware.api [None req-1ece652d-e3a1-4b82-be99-c1747ddf14f1 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285081, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.209127] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f1666685-2914-415b-a4ae-382918967cb8 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lock "0fdb21d1-4111-4ff3-bdc0-e2598298a9a4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.110s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 883.209424] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f1666685-2914-415b-a4ae-382918967cb8 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquiring lock "0fdb21d1-4111-4ff3-bdc0-e2598298a9a4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 883.209695] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f1666685-2914-415b-a4ae-382918967cb8 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lock "0fdb21d1-4111-4ff3-bdc0-e2598298a9a4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 883.209940] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f1666685-2914-415b-a4ae-382918967cb8 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lock "0fdb21d1-4111-4ff3-bdc0-e2598298a9a4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 883.212516] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49cb4bf3-485f-4871-ba6b-1499eb6820f3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.215466] env[63088]: INFO nova.compute.manager [None req-f1666685-2914-415b-a4ae-382918967cb8 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 0fdb21d1-4111-4ff3-bdc0-e2598298a9a4] Terminating instance [ 883.217616] env[63088]: DEBUG nova.compute.manager [None req-f1666685-2914-415b-a4ae-382918967cb8 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 0fdb21d1-4111-4ff3-bdc0-e2598298a9a4] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 883.217874] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-f1666685-2914-415b-a4ae-382918967cb8 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 0fdb21d1-4111-4ff3-bdc0-e2598298a9a4] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 883.218632] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f2a98c8-aee5-484f-9beb-cf8cffe112b2 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.225078] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4210b64-f4c0-4cff-bf35-5c55b6713ddb {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.230847] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1666685-2914-415b-a4ae-382918967cb8 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 0fdb21d1-4111-4ff3-bdc0-e2598298a9a4] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 883.231496] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fcb3ee70-43f5-4826-b1c9-8d62e701b7ff {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.243797] env[63088]: DEBUG nova.compute.provider_tree [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 883.246630] env[63088]: DEBUG oslo_vmware.api [None req-f1666685-2914-415b-a4ae-382918967cb8 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Waiting for the task: (returnval){ [ 883.246630] env[63088]: value = "task-1285084" [ 883.246630] env[63088]: _type = "Task" [ 883.246630] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.256926] env[63088]: DEBUG oslo_vmware.api [None req-f1666685-2914-415b-a4ae-382918967cb8 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285084, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.408225] env[63088]: DEBUG oslo_vmware.api [None req-c799bf18-c9a4-4b67-989b-6dda4cd4483c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285083, 'name': ReconfigVM_Task, 'duration_secs': 0.255723} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.408655] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-c799bf18-c9a4-4b67-989b-6dda4cd4483c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 9d5f1cde-e787-4a2c-ac49-83ec135ff51c] Reconfigured VM instance instance-00000046 to detach disk 2001 {{(pid=63088) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 883.413360] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-95a80292-2677-4cea-a0ea-fa1dbf6ab5d5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.431625] env[63088]: DEBUG oslo_vmware.api [None req-c799bf18-c9a4-4b67-989b-6dda4cd4483c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 883.431625] env[63088]: value = "task-1285085" [ 883.431625] env[63088]: _type = "Task" [ 883.431625] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.440537] env[63088]: DEBUG oslo_vmware.api [None req-c799bf18-c9a4-4b67-989b-6dda4cd4483c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285085, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.502131] env[63088]: DEBUG oslo_concurrency.lockutils [req-0e5dd796-c09c-42e5-957a-fd7d271d0e61 req-4c78697f-0b8d-4e42-8e6a-dbc49a97328a service nova] Releasing lock "refresh_cache-667fd5e9-5fe4-41e1-9d8a-896c6e5c6286" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 883.502725] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Acquired lock "refresh_cache-667fd5e9-5fe4-41e1-9d8a-896c6e5c6286" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.502960] env[63088]: DEBUG nova.network.neutron [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: 667fd5e9-5fe4-41e1-9d8a-896c6e5c6286] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 883.680756] env[63088]: DEBUG oslo_vmware.api [None req-1ece652d-e3a1-4b82-be99-c1747ddf14f1 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285081, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.749045] env[63088]: DEBUG nova.scheduler.client.report [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 883.763339] env[63088]: DEBUG oslo_vmware.api [None req-f1666685-2914-415b-a4ae-382918967cb8 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285084, 'name': PowerOffVM_Task, 'duration_secs': 0.191808} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.763594] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1666685-2914-415b-a4ae-382918967cb8 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 0fdb21d1-4111-4ff3-bdc0-e2598298a9a4] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 883.763764] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-f1666685-2914-415b-a4ae-382918967cb8 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 0fdb21d1-4111-4ff3-bdc0-e2598298a9a4] Unregistering the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 883.764034] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4d13558b-fcd1-4842-9833-a0e8722d4295 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.856380] env[63088]: DEBUG nova.compute.manager [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: b50db882-598c-488f-b935-34f55f655642] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 883.884340] env[63088]: DEBUG nova.virt.hardware [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 883.884641] env[63088]: DEBUG nova.virt.hardware [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 883.884821] env[63088]: DEBUG nova.virt.hardware [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 883.885081] env[63088]: DEBUG nova.virt.hardware [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 883.885287] env[63088]: DEBUG nova.virt.hardware [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 883.885548] env[63088]: DEBUG nova.virt.hardware [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 883.885722] env[63088]: DEBUG nova.virt.hardware [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 883.885921] env[63088]: DEBUG nova.virt.hardware [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 883.886157] env[63088]: DEBUG nova.virt.hardware [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 883.886352] env[63088]: DEBUG nova.virt.hardware [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 883.886599] env[63088]: DEBUG nova.virt.hardware [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 883.887889] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f85b2661-435f-4b40-b0a7-d0d195f8f02a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.897287] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ffcc28c-cc92-42bd-a935-72d003e2ce86 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.942872] env[63088]: DEBUG oslo_vmware.api [None req-c799bf18-c9a4-4b67-989b-6dda4cd4483c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285085, 'name': ReconfigVM_Task, 'duration_secs': 0.160919} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.943206] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-c799bf18-c9a4-4b67-989b-6dda4cd4483c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 9d5f1cde-e787-4a2c-ac49-83ec135ff51c] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-275890', 'volume_id': 'b674aa10-140a-4cb0-8ca0-7fa9fddf7a6d', 'name': 'volume-b674aa10-140a-4cb0-8ca0-7fa9fddf7a6d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9d5f1cde-e787-4a2c-ac49-83ec135ff51c', 'attached_at': '', 'detached_at': '', 'volume_id': 'b674aa10-140a-4cb0-8ca0-7fa9fddf7a6d', 'serial': 'b674aa10-140a-4cb0-8ca0-7fa9fddf7a6d'} {{(pid=63088) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 883.943503] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-c799bf18-c9a4-4b67-989b-6dda4cd4483c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 9d5f1cde-e787-4a2c-ac49-83ec135ff51c] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 883.944305] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b20fdfba-0cc5-461c-8d68-6560fd9d0c8c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.952699] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-c799bf18-c9a4-4b67-989b-6dda4cd4483c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 9d5f1cde-e787-4a2c-ac49-83ec135ff51c] Unregistering the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 883.952699] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cc3c411d-47d4-45d1-91f4-872229e756bb {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.986492] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-f1666685-2914-415b-a4ae-382918967cb8 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 0fdb21d1-4111-4ff3-bdc0-e2598298a9a4] Unregistered the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 883.986740] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-f1666685-2914-415b-a4ae-382918967cb8 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 0fdb21d1-4111-4ff3-bdc0-e2598298a9a4] Deleting contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 883.987023] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-f1666685-2914-415b-a4ae-382918967cb8 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Deleting the datastore file [datastore1] 0fdb21d1-4111-4ff3-bdc0-e2598298a9a4 {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 883.987343] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7244d511-169c-4b9c-89fd-0f53fdb84f66 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.995437] env[63088]: DEBUG oslo_vmware.api [None req-f1666685-2914-415b-a4ae-382918967cb8 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Waiting for the task: (returnval){ [ 883.995437] env[63088]: value = "task-1285088" [ 883.995437] env[63088]: _type = "Task" [ 883.995437] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.008263] env[63088]: DEBUG oslo_vmware.api [None req-f1666685-2914-415b-a4ae-382918967cb8 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285088, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.018823] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-c799bf18-c9a4-4b67-989b-6dda4cd4483c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 9d5f1cde-e787-4a2c-ac49-83ec135ff51c] Unregistered the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 884.019076] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-c799bf18-c9a4-4b67-989b-6dda4cd4483c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 9d5f1cde-e787-4a2c-ac49-83ec135ff51c] Deleting contents of the VM from datastore datastore2 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 884.019267] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-c799bf18-c9a4-4b67-989b-6dda4cd4483c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Deleting the datastore file [datastore2] 9d5f1cde-e787-4a2c-ac49-83ec135ff51c {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 884.019549] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-98a6fcaf-8df7-4d26-bc06-8c0af4bf3a63 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.027629] env[63088]: DEBUG oslo_vmware.api [None req-c799bf18-c9a4-4b67-989b-6dda4cd4483c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 884.027629] env[63088]: value = "task-1285089" [ 884.027629] env[63088]: _type = "Task" [ 884.027629] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.037267] env[63088]: DEBUG oslo_vmware.api [None req-c799bf18-c9a4-4b67-989b-6dda4cd4483c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285089, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.050676] env[63088]: DEBUG nova.network.neutron [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: 667fd5e9-5fe4-41e1-9d8a-896c6e5c6286] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 884.185417] env[63088]: DEBUG oslo_vmware.api [None req-1ece652d-e3a1-4b82-be99-c1747ddf14f1 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285081, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.240307] env[63088]: DEBUG nova.network.neutron [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: 667fd5e9-5fe4-41e1-9d8a-896c6e5c6286] Updating instance_info_cache with network_info: [{"id": "602aeee4-c8d8-41c6-b145-dae2eadfc9a7", "address": "fa:16:3e:93:f6:b8", "network": {"id": "4020a7f4-37ac-4fa2-a11d-e5f6def59125", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-770080631-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a421ece306c64913b7cf9b55709a49c5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c1520c99-af74-4d61-a8ae-56aef56ef4f0", "external-id": "nsx-vlan-transportzone-891", "segmentation_id": 891, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap602aeee4-c8", "ovs_interfaceid": "602aeee4-c8d8-41c6-b145-dae2eadfc9a7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 884.257640] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.447s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 884.257640] env[63088]: DEBUG nova.compute.manager [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: 2964db35-7357-40a7-b4e6-7e2595549f5b] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 884.261414] env[63088]: DEBUG oslo_concurrency.lockutils [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 20.768s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 884.342934] env[63088]: DEBUG nova.compute.manager [req-92e56880-3eda-44af-8f81-9ea4bcb1af96 req-05da6a08-5876-48ae-aedd-8be3bef9a28a service nova] [instance: b50db882-598c-488f-b935-34f55f655642] Received event network-vif-plugged-585051d5-6974-4abd-864e-38967371f048 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 884.343290] env[63088]: DEBUG oslo_concurrency.lockutils [req-92e56880-3eda-44af-8f81-9ea4bcb1af96 req-05da6a08-5876-48ae-aedd-8be3bef9a28a service nova] Acquiring lock "b50db882-598c-488f-b935-34f55f655642-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.343609] env[63088]: DEBUG oslo_concurrency.lockutils [req-92e56880-3eda-44af-8f81-9ea4bcb1af96 req-05da6a08-5876-48ae-aedd-8be3bef9a28a service nova] Lock "b50db882-598c-488f-b935-34f55f655642-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 884.343875] env[63088]: DEBUG oslo_concurrency.lockutils [req-92e56880-3eda-44af-8f81-9ea4bcb1af96 req-05da6a08-5876-48ae-aedd-8be3bef9a28a service nova] Lock "b50db882-598c-488f-b935-34f55f655642-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 884.344150] env[63088]: DEBUG nova.compute.manager [req-92e56880-3eda-44af-8f81-9ea4bcb1af96 req-05da6a08-5876-48ae-aedd-8be3bef9a28a service nova] [instance: b50db882-598c-488f-b935-34f55f655642] No waiting events found dispatching network-vif-plugged-585051d5-6974-4abd-864e-38967371f048 {{(pid=63088) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 884.344411] env[63088]: WARNING nova.compute.manager [req-92e56880-3eda-44af-8f81-9ea4bcb1af96 req-05da6a08-5876-48ae-aedd-8be3bef9a28a service nova] [instance: b50db882-598c-488f-b935-34f55f655642] Received unexpected event network-vif-plugged-585051d5-6974-4abd-864e-38967371f048 for instance with vm_state building and task_state spawning. [ 884.430530] env[63088]: DEBUG nova.network.neutron [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: b50db882-598c-488f-b935-34f55f655642] Successfully updated port: 585051d5-6974-4abd-864e-38967371f048 {{(pid=63088) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 884.506626] env[63088]: DEBUG oslo_vmware.api [None req-f1666685-2914-415b-a4ae-382918967cb8 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285088, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.139772} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.506882] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-f1666685-2914-415b-a4ae-382918967cb8 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 884.507045] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-f1666685-2914-415b-a4ae-382918967cb8 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 0fdb21d1-4111-4ff3-bdc0-e2598298a9a4] Deleted contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 884.507240] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-f1666685-2914-415b-a4ae-382918967cb8 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 0fdb21d1-4111-4ff3-bdc0-e2598298a9a4] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 884.507435] env[63088]: INFO nova.compute.manager [None req-f1666685-2914-415b-a4ae-382918967cb8 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 0fdb21d1-4111-4ff3-bdc0-e2598298a9a4] Took 1.29 seconds to destroy the instance on the hypervisor. [ 884.507689] env[63088]: DEBUG oslo.service.loopingcall [None req-f1666685-2914-415b-a4ae-382918967cb8 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 884.508273] env[63088]: DEBUG nova.compute.manager [-] [instance: 0fdb21d1-4111-4ff3-bdc0-e2598298a9a4] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 884.508398] env[63088]: DEBUG nova.network.neutron [-] [instance: 0fdb21d1-4111-4ff3-bdc0-e2598298a9a4] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 884.537902] env[63088]: DEBUG oslo_vmware.api [None req-c799bf18-c9a4-4b67-989b-6dda4cd4483c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285089, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.151699} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.538179] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-c799bf18-c9a4-4b67-989b-6dda4cd4483c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 884.538366] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-c799bf18-c9a4-4b67-989b-6dda4cd4483c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 9d5f1cde-e787-4a2c-ac49-83ec135ff51c] Deleted contents of the VM from datastore datastore2 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 884.538568] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-c799bf18-c9a4-4b67-989b-6dda4cd4483c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 9d5f1cde-e787-4a2c-ac49-83ec135ff51c] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 884.538773] env[63088]: INFO nova.compute.manager [None req-c799bf18-c9a4-4b67-989b-6dda4cd4483c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 9d5f1cde-e787-4a2c-ac49-83ec135ff51c] Took 2.29 seconds to destroy the instance on the hypervisor. [ 884.539031] env[63088]: DEBUG oslo.service.loopingcall [None req-c799bf18-c9a4-4b67-989b-6dda4cd4483c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 884.539231] env[63088]: DEBUG nova.compute.manager [-] [instance: 9d5f1cde-e787-4a2c-ac49-83ec135ff51c] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 884.539324] env[63088]: DEBUG nova.network.neutron [-] [instance: 9d5f1cde-e787-4a2c-ac49-83ec135ff51c] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 884.682724] env[63088]: DEBUG oslo_vmware.api [None req-1ece652d-e3a1-4b82-be99-c1747ddf14f1 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285081, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.743429] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Releasing lock "refresh_cache-667fd5e9-5fe4-41e1-9d8a-896c6e5c6286" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 884.743778] env[63088]: DEBUG nova.compute.manager [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: 667fd5e9-5fe4-41e1-9d8a-896c6e5c6286] Instance network_info: |[{"id": "602aeee4-c8d8-41c6-b145-dae2eadfc9a7", "address": "fa:16:3e:93:f6:b8", "network": {"id": "4020a7f4-37ac-4fa2-a11d-e5f6def59125", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-770080631-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a421ece306c64913b7cf9b55709a49c5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c1520c99-af74-4d61-a8ae-56aef56ef4f0", "external-id": "nsx-vlan-transportzone-891", "segmentation_id": 891, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap602aeee4-c8", "ovs_interfaceid": "602aeee4-c8d8-41c6-b145-dae2eadfc9a7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 884.744341] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: 667fd5e9-5fe4-41e1-9d8a-896c6e5c6286] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:93:f6:b8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c1520c99-af74-4d61-a8ae-56aef56ef4f0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '602aeee4-c8d8-41c6-b145-dae2eadfc9a7', 'vif_model': 'vmxnet3'}] {{(pid=63088) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 884.752186] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Creating folder: Project (a421ece306c64913b7cf9b55709a49c5). Parent ref: group-v275816. {{(pid=63088) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 884.752501] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cd81ef8f-bbe4-4677-af1b-50d7c795de38 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.767574] env[63088]: INFO nova.compute.claims [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 884.775121] env[63088]: DEBUG nova.compute.utils [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 884.775121] env[63088]: INFO nova.virt.vmwareapi.vm_util [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Created folder: Project (a421ece306c64913b7cf9b55709a49c5) in parent group-v275816. [ 884.775121] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Creating folder: Instances. Parent ref: group-v275892. {{(pid=63088) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 884.775121] env[63088]: DEBUG nova.compute.manager [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: 2964db35-7357-40a7-b4e6-7e2595549f5b] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 884.775121] env[63088]: DEBUG nova.network.neutron [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: 2964db35-7357-40a7-b4e6-7e2595549f5b] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 884.777504] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8ac663aa-949b-4283-8929-5cff3b87bf12 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.790155] env[63088]: INFO nova.virt.vmwareapi.vm_util [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Created folder: Instances in parent group-v275892. [ 884.790436] env[63088]: DEBUG oslo.service.loopingcall [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 884.790642] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 667fd5e9-5fe4-41e1-9d8a-896c6e5c6286] Creating VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 884.790964] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d3de6037-62ef-46d5-9aae-cc507027e66c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.813591] env[63088]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 884.813591] env[63088]: value = "task-1285092" [ 884.813591] env[63088]: _type = "Task" [ 884.813591] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.822931] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285092, 'name': CreateVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.824746] env[63088]: DEBUG nova.policy [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '065ecdba92ea4c5a9f6f493eefd6fa51', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a421ece306c64913b7cf9b55709a49c5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 884.933354] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Acquiring lock "refresh_cache-b50db882-598c-488f-b935-34f55f655642" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 884.933584] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Acquired lock "refresh_cache-b50db882-598c-488f-b935-34f55f655642" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 884.933833] env[63088]: DEBUG nova.network.neutron [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: b50db882-598c-488f-b935-34f55f655642] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 885.185863] env[63088]: DEBUG oslo_vmware.api [None req-1ece652d-e3a1-4b82-be99-c1747ddf14f1 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285081, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.270857] env[63088]: DEBUG nova.network.neutron [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: 2964db35-7357-40a7-b4e6-7e2595549f5b] Successfully created port: b42d0eed-48b4-4937-b222-80a994ea7602 {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 885.274680] env[63088]: DEBUG nova.compute.manager [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: 2964db35-7357-40a7-b4e6-7e2595549f5b] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 885.278669] env[63088]: INFO nova.compute.resource_tracker [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Updating resource usage from migration 02e19d97-e38c-43e6-b757-056d827d87dc [ 885.327439] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285092, 'name': CreateVM_Task, 'duration_secs': 0.419625} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.330437] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 667fd5e9-5fe4-41e1-9d8a-896c6e5c6286] Created VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 885.331518] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 885.331662] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.331995] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 885.332345] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dc6e0f91-07d8-46fb-8196-f24c4f26672d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.337994] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Waiting for the task: (returnval){ [ 885.337994] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52fa3522-ab07-f627-5c3e-6bd58eb2dc41" [ 885.337994] env[63088]: _type = "Task" [ 885.337994] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.349435] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52fa3522-ab07-f627-5c3e-6bd58eb2dc41, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.417183] env[63088]: DEBUG nova.network.neutron [-] [instance: 0fdb21d1-4111-4ff3-bdc0-e2598298a9a4] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 885.419952] env[63088]: DEBUG nova.network.neutron [-] [instance: 9d5f1cde-e787-4a2c-ac49-83ec135ff51c] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 885.465840] env[63088]: DEBUG nova.network.neutron [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: b50db882-598c-488f-b935-34f55f655642] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 885.571730] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81c9891b-54a6-4506-a4ce-46975316f8bf {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.580889] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2bdff58-efb5-4820-8252-e5c0f50df6e6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.613566] env[63088]: DEBUG nova.network.neutron [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: b50db882-598c-488f-b935-34f55f655642] Updating instance_info_cache with network_info: [{"id": "585051d5-6974-4abd-864e-38967371f048", "address": "fa:16:3e:0f:58:73", "network": {"id": "4020a7f4-37ac-4fa2-a11d-e5f6def59125", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-770080631-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a421ece306c64913b7cf9b55709a49c5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c1520c99-af74-4d61-a8ae-56aef56ef4f0", "external-id": "nsx-vlan-transportzone-891", "segmentation_id": 891, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap585051d5-69", "ovs_interfaceid": "585051d5-6974-4abd-864e-38967371f048", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 885.615325] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-797c5181-9c59-4291-ae0a-7541a2a6ac87 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.624115] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9da3de9c-4e8b-4bf9-86b5-f17a495143d1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.640175] env[63088]: DEBUG nova.compute.provider_tree [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 885.685822] env[63088]: DEBUG oslo_vmware.api [None req-1ece652d-e3a1-4b82-be99-c1747ddf14f1 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285081, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.849517] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52fa3522-ab07-f627-5c3e-6bd58eb2dc41, 'name': SearchDatastore_Task, 'duration_secs': 0.010697} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.849831] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.850167] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: 667fd5e9-5fe4-41e1-9d8a-896c6e5c6286] Processing image 1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 885.850407] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 885.850556] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.850735] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 885.851049] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-20664123-54b2-4d59-b176-d4a774310808 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.860310] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 885.860500] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 885.861267] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1eb34c47-d790-4311-9765-b2100341916e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.866768] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Waiting for the task: (returnval){ [ 885.866768] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52943539-2bf5-1b3f-fb5c-26a8f0ce49b4" [ 885.866768] env[63088]: _type = "Task" [ 885.866768] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.874847] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52943539-2bf5-1b3f-fb5c-26a8f0ce49b4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.920100] env[63088]: INFO nova.compute.manager [-] [instance: 0fdb21d1-4111-4ff3-bdc0-e2598298a9a4] Took 1.41 seconds to deallocate network for instance. [ 885.922141] env[63088]: INFO nova.compute.manager [-] [instance: 9d5f1cde-e787-4a2c-ac49-83ec135ff51c] Took 1.38 seconds to deallocate network for instance. [ 886.119025] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Releasing lock "refresh_cache-b50db882-598c-488f-b935-34f55f655642" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 886.119416] env[63088]: DEBUG nova.compute.manager [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: b50db882-598c-488f-b935-34f55f655642] Instance network_info: |[{"id": "585051d5-6974-4abd-864e-38967371f048", "address": "fa:16:3e:0f:58:73", "network": {"id": "4020a7f4-37ac-4fa2-a11d-e5f6def59125", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-770080631-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a421ece306c64913b7cf9b55709a49c5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c1520c99-af74-4d61-a8ae-56aef56ef4f0", "external-id": "nsx-vlan-transportzone-891", "segmentation_id": 891, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap585051d5-69", "ovs_interfaceid": "585051d5-6974-4abd-864e-38967371f048", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 886.119916] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: b50db882-598c-488f-b935-34f55f655642] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0f:58:73', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c1520c99-af74-4d61-a8ae-56aef56ef4f0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '585051d5-6974-4abd-864e-38967371f048', 'vif_model': 'vmxnet3'}] {{(pid=63088) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 886.128546] env[63088]: DEBUG oslo.service.loopingcall [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 886.128810] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b50db882-598c-488f-b935-34f55f655642] Creating VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 886.129053] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d9dbb86c-a976-4362-9968-cb9296eb097b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.144644] env[63088]: DEBUG nova.scheduler.client.report [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 886.156885] env[63088]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 886.156885] env[63088]: value = "task-1285093" [ 886.156885] env[63088]: _type = "Task" [ 886.156885] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.167080] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285093, 'name': CreateVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.187380] env[63088]: DEBUG oslo_vmware.api [None req-1ece652d-e3a1-4b82-be99-c1747ddf14f1 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285081, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.289195] env[63088]: DEBUG nova.compute.manager [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: 2964db35-7357-40a7-b4e6-7e2595549f5b] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 886.319922] env[63088]: DEBUG nova.virt.hardware [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 886.320246] env[63088]: DEBUG nova.virt.hardware [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 886.320516] env[63088]: DEBUG nova.virt.hardware [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 886.320875] env[63088]: DEBUG nova.virt.hardware [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 886.321106] env[63088]: DEBUG nova.virt.hardware [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 886.321373] env[63088]: DEBUG nova.virt.hardware [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 886.321619] env[63088]: DEBUG nova.virt.hardware [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 886.321789] env[63088]: DEBUG nova.virt.hardware [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 886.321990] env[63088]: DEBUG nova.virt.hardware [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 886.322169] env[63088]: DEBUG nova.virt.hardware [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 886.322347] env[63088]: DEBUG nova.virt.hardware [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 886.323226] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-781e182c-aa62-426b-9fd7-7febd9e12dae {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.334578] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6184dbd7-7ea5-4b48-bfb0-a0767c0bffd2 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.380869] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52943539-2bf5-1b3f-fb5c-26a8f0ce49b4, 'name': SearchDatastore_Task, 'duration_secs': 0.010176} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.381730] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-344a7949-4220-4f7d-b72c-33fadfdec70f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.388044] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Waiting for the task: (returnval){ [ 886.388044] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52cb2208-10f7-0580-6e5e-c982c9aa1131" [ 886.388044] env[63088]: _type = "Task" [ 886.388044] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.396667] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52cb2208-10f7-0580-6e5e-c982c9aa1131, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.429517] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f1666685-2914-415b-a4ae-382918967cb8 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 886.437725] env[63088]: DEBUG nova.compute.manager [req-95208b6e-41c4-4c82-91d2-03ae3c5430d9 req-ab42caa7-8a6c-4e11-baa6-95b86c6fface service nova] [instance: b50db882-598c-488f-b935-34f55f655642] Received event network-changed-585051d5-6974-4abd-864e-38967371f048 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 886.437845] env[63088]: DEBUG nova.compute.manager [req-95208b6e-41c4-4c82-91d2-03ae3c5430d9 req-ab42caa7-8a6c-4e11-baa6-95b86c6fface service nova] [instance: b50db882-598c-488f-b935-34f55f655642] Refreshing instance network info cache due to event network-changed-585051d5-6974-4abd-864e-38967371f048. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 886.438064] env[63088]: DEBUG oslo_concurrency.lockutils [req-95208b6e-41c4-4c82-91d2-03ae3c5430d9 req-ab42caa7-8a6c-4e11-baa6-95b86c6fface service nova] Acquiring lock "refresh_cache-b50db882-598c-488f-b935-34f55f655642" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 886.438216] env[63088]: DEBUG oslo_concurrency.lockutils [req-95208b6e-41c4-4c82-91d2-03ae3c5430d9 req-ab42caa7-8a6c-4e11-baa6-95b86c6fface service nova] Acquired lock "refresh_cache-b50db882-598c-488f-b935-34f55f655642" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 886.438380] env[63088]: DEBUG nova.network.neutron [req-95208b6e-41c4-4c82-91d2-03ae3c5430d9 req-ab42caa7-8a6c-4e11-baa6-95b86c6fface service nova] [instance: b50db882-598c-488f-b935-34f55f655642] Refreshing network info cache for port 585051d5-6974-4abd-864e-38967371f048 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 886.476330] env[63088]: INFO nova.compute.manager [None req-c799bf18-c9a4-4b67-989b-6dda4cd4483c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 9d5f1cde-e787-4a2c-ac49-83ec135ff51c] Took 0.55 seconds to detach 1 volumes for instance. [ 886.653227] env[63088]: DEBUG oslo_concurrency.lockutils [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.389s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.653227] env[63088]: INFO nova.compute.manager [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Migrating [ 886.653227] env[63088]: DEBUG oslo_concurrency.lockutils [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Acquiring lock "compute-rpcapi-router" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 886.653227] env[63088]: DEBUG oslo_concurrency.lockutils [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Acquired lock "compute-rpcapi-router" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 886.653227] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 19.258s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 886.653227] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.653841] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63088) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 886.655032] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.926s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 886.658138] env[63088]: INFO nova.compute.claims [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 254db932-35f1-42e3-9207-cd886efd65a3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 886.662197] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-200c4dda-8836-4af0-8d9c-6a349edb6375 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.682030] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b42e9562-adbe-42fd-87f1-adeec69de49b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.687293] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285093, 'name': CreateVM_Task} progress is 99%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.696281] env[63088]: DEBUG nova.compute.manager [req-5a3c2ab1-190f-4344-bb45-edbee16c8e32 req-b2a348fc-b253-4ba7-afb6-fab745924186 service nova] [instance: 2964db35-7357-40a7-b4e6-7e2595549f5b] Received event network-vif-plugged-b42d0eed-48b4-4937-b222-80a994ea7602 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 886.696281] env[63088]: DEBUG oslo_concurrency.lockutils [req-5a3c2ab1-190f-4344-bb45-edbee16c8e32 req-b2a348fc-b253-4ba7-afb6-fab745924186 service nova] Acquiring lock "2964db35-7357-40a7-b4e6-7e2595549f5b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 886.696547] env[63088]: DEBUG oslo_concurrency.lockutils [req-5a3c2ab1-190f-4344-bb45-edbee16c8e32 req-b2a348fc-b253-4ba7-afb6-fab745924186 service nova] Lock "2964db35-7357-40a7-b4e6-7e2595549f5b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 886.697058] env[63088]: DEBUG oslo_concurrency.lockutils [req-5a3c2ab1-190f-4344-bb45-edbee16c8e32 req-b2a348fc-b253-4ba7-afb6-fab745924186 service nova] Lock "2964db35-7357-40a7-b4e6-7e2595549f5b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.697477] env[63088]: DEBUG nova.compute.manager [req-5a3c2ab1-190f-4344-bb45-edbee16c8e32 req-b2a348fc-b253-4ba7-afb6-fab745924186 service nova] [instance: 2964db35-7357-40a7-b4e6-7e2595549f5b] No waiting events found dispatching network-vif-plugged-b42d0eed-48b4-4937-b222-80a994ea7602 {{(pid=63088) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 886.698298] env[63088]: WARNING nova.compute.manager [req-5a3c2ab1-190f-4344-bb45-edbee16c8e32 req-b2a348fc-b253-4ba7-afb6-fab745924186 service nova] [instance: 2964db35-7357-40a7-b4e6-7e2595549f5b] Received unexpected event network-vif-plugged-b42d0eed-48b4-4937-b222-80a994ea7602 for instance with vm_state building and task_state spawning. [ 886.710794] env[63088]: DEBUG oslo_vmware.api [None req-1ece652d-e3a1-4b82-be99-c1747ddf14f1 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285081, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.713470] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b10060b-6f8f-4bd9-8743-e2aaa8eeabcd {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.722050] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9d78f58-5895-4a2c-a036-e43b0fa3c59e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.763431] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180774MB free_disk=140GB free_vcpus=48 pci_devices=None {{(pid=63088) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 886.763648] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 886.764755] env[63088]: DEBUG nova.network.neutron [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: 2964db35-7357-40a7-b4e6-7e2595549f5b] Successfully updated port: b42d0eed-48b4-4937-b222-80a994ea7602 {{(pid=63088) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 886.770087] env[63088]: DEBUG nova.compute.manager [req-16bc6565-bce0-4b27-b91e-aed160ae6670 req-e89bf1e9-2ecb-4376-a93f-5ac316887177 service nova] [instance: 2964db35-7357-40a7-b4e6-7e2595549f5b] Received event network-changed-b42d0eed-48b4-4937-b222-80a994ea7602 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 886.770372] env[63088]: DEBUG nova.compute.manager [req-16bc6565-bce0-4b27-b91e-aed160ae6670 req-e89bf1e9-2ecb-4376-a93f-5ac316887177 service nova] [instance: 2964db35-7357-40a7-b4e6-7e2595549f5b] Refreshing instance network info cache due to event network-changed-b42d0eed-48b4-4937-b222-80a994ea7602. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 886.770642] env[63088]: DEBUG oslo_concurrency.lockutils [req-16bc6565-bce0-4b27-b91e-aed160ae6670 req-e89bf1e9-2ecb-4376-a93f-5ac316887177 service nova] Acquiring lock "refresh_cache-2964db35-7357-40a7-b4e6-7e2595549f5b" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 886.770850] env[63088]: DEBUG oslo_concurrency.lockutils [req-16bc6565-bce0-4b27-b91e-aed160ae6670 req-e89bf1e9-2ecb-4376-a93f-5ac316887177 service nova] Acquired lock "refresh_cache-2964db35-7357-40a7-b4e6-7e2595549f5b" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 886.771117] env[63088]: DEBUG nova.network.neutron [req-16bc6565-bce0-4b27-b91e-aed160ae6670 req-e89bf1e9-2ecb-4376-a93f-5ac316887177 service nova] [instance: 2964db35-7357-40a7-b4e6-7e2595549f5b] Refreshing network info cache for port b42d0eed-48b4-4937-b222-80a994ea7602 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 886.901032] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52cb2208-10f7-0580-6e5e-c982c9aa1131, 'name': SearchDatastore_Task, 'duration_secs': 0.009844} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.901474] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 886.901906] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] 667fd5e9-5fe4-41e1-9d8a-896c6e5c6286/667fd5e9-5fe4-41e1-9d8a-896c6e5c6286.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 886.902251] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1e522d0f-ec9a-4857-b412-8c0b602b1b94 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.912315] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Waiting for the task: (returnval){ [ 886.912315] env[63088]: value = "task-1285094" [ 886.912315] env[63088]: _type = "Task" [ 886.912315] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.922837] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Task: {'id': task-1285094, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.984311] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c799bf18-c9a4-4b67-989b-6dda4cd4483c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.161996] env[63088]: INFO nova.compute.rpcapi [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Automatically selected compute RPC version 6.3 from minimum service version 67 [ 887.162641] env[63088]: DEBUG oslo_concurrency.lockutils [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Releasing lock "compute-rpcapi-router" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 887.198495] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285093, 'name': CreateVM_Task, 'duration_secs': 0.591689} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.198495] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b50db882-598c-488f-b935-34f55f655642] Created VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 887.199776] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 887.200028] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.200610] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 887.205069] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c8b456b0-417e-45de-8224-7cc0ae5e81b8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.207080] env[63088]: DEBUG oslo_vmware.api [None req-1ece652d-e3a1-4b82-be99-c1747ddf14f1 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285081, 'name': ReconfigVM_Task} progress is 18%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.212250] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Waiting for the task: (returnval){ [ 887.212250] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]524e5f4e-bcc3-943c-44e7-12b523f2f13a" [ 887.212250] env[63088]: _type = "Task" [ 887.212250] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.221122] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]524e5f4e-bcc3-943c-44e7-12b523f2f13a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.222593] env[63088]: DEBUG nova.network.neutron [req-95208b6e-41c4-4c82-91d2-03ae3c5430d9 req-ab42caa7-8a6c-4e11-baa6-95b86c6fface service nova] [instance: b50db882-598c-488f-b935-34f55f655642] Updated VIF entry in instance network info cache for port 585051d5-6974-4abd-864e-38967371f048. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 887.222593] env[63088]: DEBUG nova.network.neutron [req-95208b6e-41c4-4c82-91d2-03ae3c5430d9 req-ab42caa7-8a6c-4e11-baa6-95b86c6fface service nova] [instance: b50db882-598c-488f-b935-34f55f655642] Updating instance_info_cache with network_info: [{"id": "585051d5-6974-4abd-864e-38967371f048", "address": "fa:16:3e:0f:58:73", "network": {"id": "4020a7f4-37ac-4fa2-a11d-e5f6def59125", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-770080631-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a421ece306c64913b7cf9b55709a49c5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c1520c99-af74-4d61-a8ae-56aef56ef4f0", "external-id": "nsx-vlan-transportzone-891", "segmentation_id": 891, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap585051d5-69", "ovs_interfaceid": "585051d5-6974-4abd-864e-38967371f048", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 887.267594] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Acquiring lock "refresh_cache-2964db35-7357-40a7-b4e6-7e2595549f5b" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 887.322058] env[63088]: DEBUG nova.network.neutron [req-16bc6565-bce0-4b27-b91e-aed160ae6670 req-e89bf1e9-2ecb-4376-a93f-5ac316887177 service nova] [instance: 2964db35-7357-40a7-b4e6-7e2595549f5b] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 887.423681] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Task: {'id': task-1285094, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.433329] env[63088]: DEBUG nova.network.neutron [req-16bc6565-bce0-4b27-b91e-aed160ae6670 req-e89bf1e9-2ecb-4376-a93f-5ac316887177 service nova] [instance: 2964db35-7357-40a7-b4e6-7e2595549f5b] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 887.686349] env[63088]: DEBUG oslo_concurrency.lockutils [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Acquiring lock "refresh_cache-e6b0ce08-d67f-458f-92f5-1e904d03a5ef" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 887.686671] env[63088]: DEBUG oslo_concurrency.lockutils [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Acquired lock "refresh_cache-e6b0ce08-d67f-458f-92f5-1e904d03a5ef" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.686881] env[63088]: DEBUG nova.network.neutron [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 887.697444] env[63088]: DEBUG oslo_vmware.api [None req-1ece652d-e3a1-4b82-be99-c1747ddf14f1 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285081, 'name': ReconfigVM_Task, 'duration_secs': 5.854715} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.699791] env[63088]: DEBUG oslo_concurrency.lockutils [None req-1ece652d-e3a1-4b82-be99-c1747ddf14f1 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Releasing lock "4971b24c-6710-4f50-9846-727dad264b1f" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 887.700040] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-1ece652d-e3a1-4b82-be99-c1747ddf14f1 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Reconfigured VM to detach interface {{(pid=63088) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1984}} [ 887.722408] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]524e5f4e-bcc3-943c-44e7-12b523f2f13a, 'name': SearchDatastore_Task, 'duration_secs': 0.057738} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.724643] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 887.724877] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: b50db882-598c-488f-b935-34f55f655642] Processing image 1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 887.725120] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 887.725269] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.725448] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 887.725896] env[63088]: DEBUG oslo_concurrency.lockutils [req-95208b6e-41c4-4c82-91d2-03ae3c5430d9 req-ab42caa7-8a6c-4e11-baa6-95b86c6fface service nova] Releasing lock "refresh_cache-b50db882-598c-488f-b935-34f55f655642" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 887.726128] env[63088]: DEBUG nova.compute.manager [req-95208b6e-41c4-4c82-91d2-03ae3c5430d9 req-ab42caa7-8a6c-4e11-baa6-95b86c6fface service nova] [instance: 0fdb21d1-4111-4ff3-bdc0-e2598298a9a4] Received event network-vif-deleted-94b61c5f-77d5-4f09-9e84-71c70bde1fcb {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 887.726376] env[63088]: DEBUG nova.compute.manager [req-95208b6e-41c4-4c82-91d2-03ae3c5430d9 req-ab42caa7-8a6c-4e11-baa6-95b86c6fface service nova] [instance: 9d5f1cde-e787-4a2c-ac49-83ec135ff51c] Received event network-vif-deleted-07e1b344-3f50-473d-a443-caca3ba6c0ea {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 887.727027] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9811784a-dcaa-4f48-b4ec-dcae9d21b40a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.735819] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 887.735945] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 887.738812] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cf3e149c-4ff4-4756-afbc-3037903a8fe8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.743998] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Waiting for the task: (returnval){ [ 887.743998] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52d9b386-f8d4-296f-6cf8-598f75e8eabd" [ 887.743998] env[63088]: _type = "Task" [ 887.743998] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.751915] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52d9b386-f8d4-296f-6cf8-598f75e8eabd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.756684] env[63088]: DEBUG oslo_concurrency.lockutils [None req-12a9b2e0-558c-4b94-9bc3-4eae7efb7c31 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquiring lock "7b6aadb7-e34b-42b7-b69f-370434f5b665" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.756933] env[63088]: DEBUG oslo_concurrency.lockutils [None req-12a9b2e0-558c-4b94-9bc3-4eae7efb7c31 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lock "7b6aadb7-e34b-42b7-b69f-370434f5b665" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.757141] env[63088]: DEBUG nova.compute.manager [None req-12a9b2e0-558c-4b94-9bc3-4eae7efb7c31 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 887.757915] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e321f1b-a66b-4337-b19b-87d5c4027355 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.763920] env[63088]: DEBUG nova.compute.manager [None req-12a9b2e0-558c-4b94-9bc3-4eae7efb7c31 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=63088) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 887.764495] env[63088]: DEBUG nova.objects.instance [None req-12a9b2e0-558c-4b94-9bc3-4eae7efb7c31 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lazy-loading 'flavor' on Instance uuid 7b6aadb7-e34b-42b7-b69f-370434f5b665 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 887.923104] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Task: {'id': task-1285094, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.526513} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.925504] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] 667fd5e9-5fe4-41e1-9d8a-896c6e5c6286/667fd5e9-5fe4-41e1-9d8a-896c6e5c6286.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 887.925730] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: 667fd5e9-5fe4-41e1-9d8a-896c6e5c6286] Extending root virtual disk to 1048576 {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 887.926170] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-003b785c-2bea-4074-b05c-bd51e06a956e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.933042] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Waiting for the task: (returnval){ [ 887.933042] env[63088]: value = "task-1285095" [ 887.933042] env[63088]: _type = "Task" [ 887.933042] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.936300] env[63088]: DEBUG oslo_concurrency.lockutils [req-16bc6565-bce0-4b27-b91e-aed160ae6670 req-e89bf1e9-2ecb-4376-a93f-5ac316887177 service nova] Releasing lock "refresh_cache-2964db35-7357-40a7-b4e6-7e2595549f5b" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 887.937614] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Acquired lock "refresh_cache-2964db35-7357-40a7-b4e6-7e2595549f5b" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.937771] env[63088]: DEBUG nova.network.neutron [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: 2964db35-7357-40a7-b4e6-7e2595549f5b] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 887.939654] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b1fff28-e397-415a-b291-a4f6ae4f8518 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.945552] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Task: {'id': task-1285095, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.950655] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ed11870-a1f9-4ef7-9812-8832a920151d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.982748] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0703c2a1-ba2c-4878-bccd-327c16bf276d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.990127] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d65af5fa-54b5-4a20-b3df-3cbe1e98bc3b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.006845] env[63088]: DEBUG nova.compute.provider_tree [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 888.254704] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52d9b386-f8d4-296f-6cf8-598f75e8eabd, 'name': SearchDatastore_Task, 'duration_secs': 0.008884} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.255522] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3901e99c-5ce7-4a59-8c2a-fc01dfad6b9b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.260787] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Waiting for the task: (returnval){ [ 888.260787] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]525df628-7510-4c45-0130-3181e40c1e79" [ 888.260787] env[63088]: _type = "Task" [ 888.260787] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.272127] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]525df628-7510-4c45-0130-3181e40c1e79, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.272707] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-12a9b2e0-558c-4b94-9bc3-4eae7efb7c31 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 888.272933] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a04cfa4a-df13-4217-b825-52e0dc3ac1aa {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.279383] env[63088]: DEBUG oslo_vmware.api [None req-12a9b2e0-558c-4b94-9bc3-4eae7efb7c31 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for the task: (returnval){ [ 888.279383] env[63088]: value = "task-1285096" [ 888.279383] env[63088]: _type = "Task" [ 888.279383] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.289579] env[63088]: DEBUG oslo_vmware.api [None req-12a9b2e0-558c-4b94-9bc3-4eae7efb7c31 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285096, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.442552] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Task: {'id': task-1285095, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.278713} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.444820] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: 667fd5e9-5fe4-41e1-9d8a-896c6e5c6286] Extended root virtual disk {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 888.445854] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec13d1b8-4426-483b-ac55-4b6653864f61 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.471531] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: 667fd5e9-5fe4-41e1-9d8a-896c6e5c6286] Reconfiguring VM instance instance-00000049 to attach disk [datastore1] 667fd5e9-5fe4-41e1-9d8a-896c6e5c6286/667fd5e9-5fe4-41e1-9d8a-896c6e5c6286.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 888.472230] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-babff37a-d493-48d8-981a-098a8233d0d5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.489260] env[63088]: DEBUG nova.network.neutron [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Updating instance_info_cache with network_info: [{"id": "63d8dbbe-4550-4d8d-8e85-9a9eadc81d82", "address": "fa:16:3e:d1:72:56", "network": {"id": "ebeefbc2-043c-4e76-aeaf-4f9113869855", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.60", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "bfb6bc5062fc4d51af1d3c577659be9a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271fe7a0-dfd7-409b-920a-cf83ef1a86a3", "external-id": "nsx-vlan-transportzone-728", "segmentation_id": 728, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap63d8dbbe-45", "ovs_interfaceid": "63d8dbbe-4550-4d8d-8e85-9a9eadc81d82", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 888.492269] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Waiting for the task: (returnval){ [ 888.492269] env[63088]: value = "task-1285097" [ 888.492269] env[63088]: _type = "Task" [ 888.492269] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.499902] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Task: {'id': task-1285097, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.510126] env[63088]: DEBUG nova.scheduler.client.report [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 888.514087] env[63088]: DEBUG nova.network.neutron [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: 2964db35-7357-40a7-b4e6-7e2595549f5b] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 888.711671] env[63088]: DEBUG nova.network.neutron [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: 2964db35-7357-40a7-b4e6-7e2595549f5b] Updating instance_info_cache with network_info: [{"id": "b42d0eed-48b4-4937-b222-80a994ea7602", "address": "fa:16:3e:49:50:ea", "network": {"id": "4020a7f4-37ac-4fa2-a11d-e5f6def59125", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-770080631-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a421ece306c64913b7cf9b55709a49c5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c1520c99-af74-4d61-a8ae-56aef56ef4f0", "external-id": "nsx-vlan-transportzone-891", "segmentation_id": 891, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb42d0eed-48", "ovs_interfaceid": "b42d0eed-48b4-4937-b222-80a994ea7602", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 888.778013] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]525df628-7510-4c45-0130-3181e40c1e79, 'name': SearchDatastore_Task, 'duration_secs': 0.033032} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.778494] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 888.778749] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] b50db882-598c-488f-b935-34f55f655642/b50db882-598c-488f-b935-34f55f655642.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 888.779020] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c6d2ebd7-c373-482f-920b-e93a87c49ff0 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.797708] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Waiting for the task: (returnval){ [ 888.797708] env[63088]: value = "task-1285098" [ 888.797708] env[63088]: _type = "Task" [ 888.797708] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.797708] env[63088]: DEBUG oslo_vmware.api [None req-12a9b2e0-558c-4b94-9bc3-4eae7efb7c31 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285096, 'name': PowerOffVM_Task, 'duration_secs': 0.228992} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.798575] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-12a9b2e0-558c-4b94-9bc3-4eae7efb7c31 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 888.798575] env[63088]: DEBUG nova.compute.manager [None req-12a9b2e0-558c-4b94-9bc3-4eae7efb7c31 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 888.803188] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e408fb95-03f0-4472-8f1e-a9497713e1f7 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.806466] env[63088]: DEBUG nova.compute.manager [req-ecb54d06-7451-4f48-9fcf-32c633776cb3 req-6d9c3d35-0d97-4c30-a811-7315cafc0f3a service nova] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Received event network-vif-deleted-81932288-bc32-47c4-a3b2-02597cb6bf81 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 888.806657] env[63088]: INFO nova.compute.manager [req-ecb54d06-7451-4f48-9fcf-32c633776cb3 req-6d9c3d35-0d97-4c30-a811-7315cafc0f3a service nova] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Neutron deleted interface 81932288-bc32-47c4-a3b2-02597cb6bf81; detaching it from the instance and deleting it from the info cache [ 888.806938] env[63088]: DEBUG nova.network.neutron [req-ecb54d06-7451-4f48-9fcf-32c633776cb3 req-6d9c3d35-0d97-4c30-a811-7315cafc0f3a service nova] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Updating instance_info_cache with network_info: [{"id": "d5a10951-ae2e-48f5-b0ca-b1f144e5fe3a", "address": "fa:16:3e:95:96:0c", "network": {"id": "1174a284-d3ca-4f9e-aa81-13ee9a693e55", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1994276040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.166", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa3d24a1a6c0430985fd80365d986ee1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd5a10951-ae", "ovs_interfaceid": "d5a10951-ae2e-48f5-b0ca-b1f144e5fe3a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "592cc24a-92a9-4bd3-bb9a-a0877ac0db04", "address": "fa:16:3e:fe:d7:1a", "network": {"id": "1174a284-d3ca-4f9e-aa81-13ee9a693e55", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1994276040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa3d24a1a6c0430985fd80365d986ee1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap592cc24a-92", "ovs_interfaceid": "592cc24a-92a9-4bd3-bb9a-a0877ac0db04", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 888.908019] env[63088]: DEBUG oslo_concurrency.lockutils [None req-1ece652d-e3a1-4b82-be99-c1747ddf14f1 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquiring lock "refresh_cache-4971b24c-6710-4f50-9846-727dad264b1f" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.908186] env[63088]: DEBUG oslo_concurrency.lockutils [None req-1ece652d-e3a1-4b82-be99-c1747ddf14f1 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquired lock "refresh_cache-4971b24c-6710-4f50-9846-727dad264b1f" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.908364] env[63088]: DEBUG nova.network.neutron [None req-1ece652d-e3a1-4b82-be99-c1747ddf14f1 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 888.992251] env[63088]: DEBUG oslo_concurrency.lockutils [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Releasing lock "refresh_cache-e6b0ce08-d67f-458f-92f5-1e904d03a5ef" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 889.004334] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Task: {'id': task-1285097, 'name': ReconfigVM_Task, 'duration_secs': 0.272781} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.004540] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: 667fd5e9-5fe4-41e1-9d8a-896c6e5c6286] Reconfigured VM instance instance-00000049 to attach disk [datastore1] 667fd5e9-5fe4-41e1-9d8a-896c6e5c6286/667fd5e9-5fe4-41e1-9d8a-896c6e5c6286.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 889.005442] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fc4f439f-953b-4f6d-b2cc-810075db09d8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.013203] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Waiting for the task: (returnval){ [ 889.013203] env[63088]: value = "task-1285099" [ 889.013203] env[63088]: _type = "Task" [ 889.013203] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.017267] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.363s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.017857] env[63088]: DEBUG nova.compute.manager [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 254db932-35f1-42e3-9207-cd886efd65a3] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 889.024124] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4ea5fd7b-e84b-41ba-9be2-088f6a9cab78 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.013s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.024381] env[63088]: DEBUG nova.objects.instance [None req-4ea5fd7b-e84b-41ba-9be2-088f6a9cab78 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Lazy-loading 'resources' on Instance uuid 8b564e29-96cf-4abf-963d-142b413fb464 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 889.025512] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Task: {'id': task-1285099, 'name': Rename_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.215011] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Releasing lock "refresh_cache-2964db35-7357-40a7-b4e6-7e2595549f5b" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 889.215512] env[63088]: DEBUG nova.compute.manager [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: 2964db35-7357-40a7-b4e6-7e2595549f5b] Instance network_info: |[{"id": "b42d0eed-48b4-4937-b222-80a994ea7602", "address": "fa:16:3e:49:50:ea", "network": {"id": "4020a7f4-37ac-4fa2-a11d-e5f6def59125", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-770080631-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a421ece306c64913b7cf9b55709a49c5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c1520c99-af74-4d61-a8ae-56aef56ef4f0", "external-id": "nsx-vlan-transportzone-891", "segmentation_id": 891, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb42d0eed-48", "ovs_interfaceid": "b42d0eed-48b4-4937-b222-80a994ea7602", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 889.216138] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: 2964db35-7357-40a7-b4e6-7e2595549f5b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:49:50:ea', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c1520c99-af74-4d61-a8ae-56aef56ef4f0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b42d0eed-48b4-4937-b222-80a994ea7602', 'vif_model': 'vmxnet3'}] {{(pid=63088) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 889.227769] env[63088]: DEBUG oslo.service.loopingcall [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 889.227769] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2964db35-7357-40a7-b4e6-7e2595549f5b] Creating VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 889.228051] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bd075ff8-f5f7-488c-bc14-4d51a02e1d26 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.248998] env[63088]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 889.248998] env[63088]: value = "task-1285100" [ 889.248998] env[63088]: _type = "Task" [ 889.248998] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.257523] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285100, 'name': CreateVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.279199] env[63088]: DEBUG oslo_concurrency.lockutils [None req-125c7f05-a006-44cd-8334-62558a6d840d tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquiring lock "4971b24c-6710-4f50-9846-727dad264b1f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.279649] env[63088]: DEBUG oslo_concurrency.lockutils [None req-125c7f05-a006-44cd-8334-62558a6d840d tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Lock "4971b24c-6710-4f50-9846-727dad264b1f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.280071] env[63088]: DEBUG oslo_concurrency.lockutils [None req-125c7f05-a006-44cd-8334-62558a6d840d tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquiring lock "4971b24c-6710-4f50-9846-727dad264b1f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.280322] env[63088]: DEBUG oslo_concurrency.lockutils [None req-125c7f05-a006-44cd-8334-62558a6d840d tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Lock "4971b24c-6710-4f50-9846-727dad264b1f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.280556] env[63088]: DEBUG oslo_concurrency.lockutils [None req-125c7f05-a006-44cd-8334-62558a6d840d tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Lock "4971b24c-6710-4f50-9846-727dad264b1f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.283201] env[63088]: INFO nova.compute.manager [None req-125c7f05-a006-44cd-8334-62558a6d840d tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Terminating instance [ 889.285383] env[63088]: DEBUG nova.compute.manager [None req-125c7f05-a006-44cd-8334-62558a6d840d tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 889.285587] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-125c7f05-a006-44cd-8334-62558a6d840d tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 889.286483] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a38764f-9d87-4563-b62b-1b5eb9220504 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.293969] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-125c7f05-a006-44cd-8334-62558a6d840d tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 889.294270] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2fe2f89d-a625-4199-b762-8f6b3c4b2d62 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.304562] env[63088]: DEBUG oslo_vmware.api [None req-125c7f05-a006-44cd-8334-62558a6d840d tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Waiting for the task: (returnval){ [ 889.304562] env[63088]: value = "task-1285101" [ 889.304562] env[63088]: _type = "Task" [ 889.304562] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.313327] env[63088]: DEBUG oslo_concurrency.lockutils [req-ecb54d06-7451-4f48-9fcf-32c633776cb3 req-6d9c3d35-0d97-4c30-a811-7315cafc0f3a service nova] Acquiring lock "4971b24c-6710-4f50-9846-727dad264b1f" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 889.313684] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Task: {'id': task-1285098, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.317177] env[63088]: DEBUG oslo_vmware.api [None req-125c7f05-a006-44cd-8334-62558a6d840d tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285101, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.324621] env[63088]: DEBUG oslo_concurrency.lockutils [None req-12a9b2e0-558c-4b94-9bc3-4eae7efb7c31 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lock "7b6aadb7-e34b-42b7-b69f-370434f5b665" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.568s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.524081] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Task: {'id': task-1285099, 'name': Rename_Task} progress is 99%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.529442] env[63088]: DEBUG nova.compute.utils [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 889.533878] env[63088]: DEBUG nova.compute.manager [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 254db932-35f1-42e3-9207-cd886efd65a3] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 889.533878] env[63088]: DEBUG nova.network.neutron [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 254db932-35f1-42e3-9207-cd886efd65a3] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 889.575771] env[63088]: DEBUG nova.policy [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a6f91fc8922c42f985f6f6338dab0708', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7bff1382b9694df08133c88a5fe783a5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 889.704422] env[63088]: INFO nova.network.neutron [None req-1ece652d-e3a1-4b82-be99-c1747ddf14f1 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Port 592cc24a-92a9-4bd3-bb9a-a0877ac0db04 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 889.704778] env[63088]: DEBUG nova.network.neutron [None req-1ece652d-e3a1-4b82-be99-c1747ddf14f1 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Updating instance_info_cache with network_info: [{"id": "d5a10951-ae2e-48f5-b0ca-b1f144e5fe3a", "address": "fa:16:3e:95:96:0c", "network": {"id": "1174a284-d3ca-4f9e-aa81-13ee9a693e55", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1994276040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.166", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa3d24a1a6c0430985fd80365d986ee1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd5a10951-ae", "ovs_interfaceid": "d5a10951-ae2e-48f5-b0ca-b1f144e5fe3a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 889.765736] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285100, 'name': CreateVM_Task} progress is 25%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.814174] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Task: {'id': task-1285098, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.525867} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.815899] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] b50db882-598c-488f-b935-34f55f655642/b50db882-598c-488f-b935-34f55f655642.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 889.816152] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: b50db882-598c-488f-b935-34f55f655642] Extending root virtual disk to 1048576 {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 889.823158] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-90956288-57f6-445c-98e7-fe04786d9721 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.825985] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7861f423-d85f-4d92-809f-e83369cab77e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.828504] env[63088]: DEBUG oslo_vmware.api [None req-125c7f05-a006-44cd-8334-62558a6d840d tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285101, 'name': PowerOffVM_Task, 'duration_secs': 0.270774} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.829423] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-125c7f05-a006-44cd-8334-62558a6d840d tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 889.830256] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-125c7f05-a006-44cd-8334-62558a6d840d tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Unregistering the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 889.830587] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3c5e0989-3fff-4d20-932f-e6bbd848ee0f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.837063] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e2d85a1-6c04-4ea5-a15f-cae3aa84666f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.841222] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Waiting for the task: (returnval){ [ 889.841222] env[63088]: value = "task-1285102" [ 889.841222] env[63088]: _type = "Task" [ 889.841222] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.875354] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba94f10d-9a60-4b61-b975-9ec1cda94afd {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.880812] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Task: {'id': task-1285102, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.886265] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16b48754-afbf-43f9-9056-d6fabe76b5ed {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.903614] env[63088]: DEBUG nova.compute.provider_tree [None req-4ea5fd7b-e84b-41ba-9be2-088f6a9cab78 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 889.976308] env[63088]: DEBUG nova.network.neutron [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 254db932-35f1-42e3-9207-cd886efd65a3] Successfully created port: d2722fc9-4b54-495f-81bb-d5b411285079 {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 889.982523] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-125c7f05-a006-44cd-8334-62558a6d840d tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Unregistered the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 889.982523] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-125c7f05-a006-44cd-8334-62558a6d840d tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Deleting contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 889.982523] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-125c7f05-a006-44cd-8334-62558a6d840d tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Deleting the datastore file [datastore1] 4971b24c-6710-4f50-9846-727dad264b1f {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 889.982523] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6b3b410a-b316-4490-8b3a-3dc17970fb67 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.988350] env[63088]: DEBUG oslo_vmware.api [None req-125c7f05-a006-44cd-8334-62558a6d840d tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Waiting for the task: (returnval){ [ 889.988350] env[63088]: value = "task-1285104" [ 889.988350] env[63088]: _type = "Task" [ 889.988350] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.996367] env[63088]: DEBUG oslo_vmware.api [None req-125c7f05-a006-44cd-8334-62558a6d840d tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285104, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.025928] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Task: {'id': task-1285099, 'name': Rename_Task} progress is 99%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.036940] env[63088]: DEBUG nova.compute.manager [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 254db932-35f1-42e3-9207-cd886efd65a3] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 890.207164] env[63088]: DEBUG oslo_concurrency.lockutils [None req-1ece652d-e3a1-4b82-be99-c1747ddf14f1 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Releasing lock "refresh_cache-4971b24c-6710-4f50-9846-727dad264b1f" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 890.260305] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285100, 'name': CreateVM_Task} progress is 99%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.331777] env[63088]: DEBUG nova.objects.instance [None req-7e0b99c7-0e13-4f0c-9208-ec4af2ac98b6 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lazy-loading 'flavor' on Instance uuid 7b6aadb7-e34b-42b7-b69f-370434f5b665 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 890.351496] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Task: {'id': task-1285102, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.08567} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.352610] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: b50db882-598c-488f-b935-34f55f655642] Extended root virtual disk {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 890.353576] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d03151b7-308d-4250-80e5-4f5b5f3060bc {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.377124] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: b50db882-598c-488f-b935-34f55f655642] Reconfiguring VM instance instance-0000004a to attach disk [datastore1] b50db882-598c-488f-b935-34f55f655642/b50db882-598c-488f-b935-34f55f655642.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 890.377863] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0200213a-7ce2-4187-9ad5-bcb797ef5e11 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.398893] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Waiting for the task: (returnval){ [ 890.398893] env[63088]: value = "task-1285105" [ 890.398893] env[63088]: _type = "Task" [ 890.398893] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.408777] env[63088]: DEBUG nova.scheduler.client.report [None req-4ea5fd7b-e84b-41ba-9be2-088f6a9cab78 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 890.412145] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Task: {'id': task-1285105, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.499623] env[63088]: DEBUG oslo_vmware.api [None req-125c7f05-a006-44cd-8334-62558a6d840d tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285104, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.216374} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.499905] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-125c7f05-a006-44cd-8334-62558a6d840d tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 890.500126] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-125c7f05-a006-44cd-8334-62558a6d840d tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Deleted contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 890.500318] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-125c7f05-a006-44cd-8334-62558a6d840d tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 890.500529] env[63088]: INFO nova.compute.manager [None req-125c7f05-a006-44cd-8334-62558a6d840d tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Took 1.21 seconds to destroy the instance on the hypervisor. [ 890.500776] env[63088]: DEBUG oslo.service.loopingcall [None req-125c7f05-a006-44cd-8334-62558a6d840d tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 890.500974] env[63088]: DEBUG nova.compute.manager [-] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 890.501077] env[63088]: DEBUG nova.network.neutron [-] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 890.511447] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08ab1e87-69d1-4346-9f4e-f1c7a803eb31 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.532258] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Updating instance 'e6b0ce08-d67f-458f-92f5-1e904d03a5ef' progress to 0 {{(pid=63088) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 890.543198] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Task: {'id': task-1285099, 'name': Rename_Task, 'duration_secs': 1.24014} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.545455] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: 667fd5e9-5fe4-41e1-9d8a-896c6e5c6286] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 890.546056] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2af7aeb8-0cd3-41de-915c-40b06766c47c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.552010] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Waiting for the task: (returnval){ [ 890.552010] env[63088]: value = "task-1285106" [ 890.552010] env[63088]: _type = "Task" [ 890.552010] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.560894] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Task: {'id': task-1285106, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.564777] env[63088]: DEBUG neutronclient.v2_0.client [-] Error message: {"NeutronError": {"type": "PortNotFound", "message": "Port 592cc24a-92a9-4bd3-bb9a-a0877ac0db04 could not be found.", "detail": ""}} {{(pid=63088) _handle_fault_response /opt/stack/data/venv/lib/python3.10/site-packages/neutronclient/v2_0/client.py:262}} [ 890.565008] env[63088]: DEBUG nova.network.neutron [-] Unable to show port 592cc24a-92a9-4bd3-bb9a-a0877ac0db04 as it no longer exists. {{(pid=63088) _unbind_ports /opt/stack/nova/nova/network/neutron.py:666}} [ 890.714105] env[63088]: DEBUG oslo_concurrency.lockutils [None req-1ece652d-e3a1-4b82-be99-c1747ddf14f1 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Lock "interface-4971b24c-6710-4f50-9846-727dad264b1f-81932288-bc32-47c4-a3b2-02597cb6bf81" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.629s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 890.761665] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285100, 'name': CreateVM_Task} progress is 99%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.836018] env[63088]: DEBUG oslo_concurrency.lockutils [None req-7e0b99c7-0e13-4f0c-9208-ec4af2ac98b6 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquiring lock "refresh_cache-7b6aadb7-e34b-42b7-b69f-370434f5b665" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 890.836289] env[63088]: DEBUG oslo_concurrency.lockutils [None req-7e0b99c7-0e13-4f0c-9208-ec4af2ac98b6 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquired lock "refresh_cache-7b6aadb7-e34b-42b7-b69f-370434f5b665" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 890.836432] env[63088]: DEBUG nova.network.neutron [None req-7e0b99c7-0e13-4f0c-9208-ec4af2ac98b6 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 890.836648] env[63088]: DEBUG nova.objects.instance [None req-7e0b99c7-0e13-4f0c-9208-ec4af2ac98b6 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lazy-loading 'info_cache' on Instance uuid 7b6aadb7-e34b-42b7-b69f-370434f5b665 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 890.908662] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Task: {'id': task-1285105, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.913898] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4ea5fd7b-e84b-41ba-9be2-088f6a9cab78 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.890s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 890.916609] env[63088]: DEBUG oslo_concurrency.lockutils [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 13.691s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 890.946772] env[63088]: INFO nova.scheduler.client.report [None req-4ea5fd7b-e84b-41ba-9be2-088f6a9cab78 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Deleted allocations for instance 8b564e29-96cf-4abf-963d-142b413fb464 [ 891.037890] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 891.038241] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-12f3e91b-0269-4f21-bebb-3559c891ec57 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.047338] env[63088]: DEBUG nova.compute.manager [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 254db932-35f1-42e3-9207-cd886efd65a3] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 891.049410] env[63088]: DEBUG oslo_vmware.api [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Waiting for the task: (returnval){ [ 891.049410] env[63088]: value = "task-1285107" [ 891.049410] env[63088]: _type = "Task" [ 891.049410] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.061681] env[63088]: DEBUG oslo_vmware.api [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Task: {'id': task-1285107, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.065471] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Task: {'id': task-1285106, 'name': PowerOnVM_Task} progress is 90%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.082415] env[63088]: DEBUG nova.virt.hardware [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 891.082686] env[63088]: DEBUG nova.virt.hardware [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 891.082847] env[63088]: DEBUG nova.virt.hardware [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 891.083050] env[63088]: DEBUG nova.virt.hardware [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 891.083213] env[63088]: DEBUG nova.virt.hardware [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 891.083365] env[63088]: DEBUG nova.virt.hardware [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 891.083577] env[63088]: DEBUG nova.virt.hardware [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 891.083745] env[63088]: DEBUG nova.virt.hardware [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 891.083916] env[63088]: DEBUG nova.virt.hardware [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 891.084192] env[63088]: DEBUG nova.virt.hardware [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 891.084391] env[63088]: DEBUG nova.virt.hardware [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 891.085260] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54473fe8-ba27-490c-a153-efcabf1aa5c6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.094673] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d718d13-12f1-4185-ad4a-47cb090b440c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.142981] env[63088]: DEBUG nova.compute.manager [req-04c9c65a-cbad-40ff-acf0-3ea530b80336 req-71b5b3fe-b289-4193-b3db-ceaf0983018f service nova] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Received event network-vif-deleted-d5a10951-ae2e-48f5-b0ca-b1f144e5fe3a {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 891.143417] env[63088]: INFO nova.compute.manager [req-04c9c65a-cbad-40ff-acf0-3ea530b80336 req-71b5b3fe-b289-4193-b3db-ceaf0983018f service nova] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Neutron deleted interface d5a10951-ae2e-48f5-b0ca-b1f144e5fe3a; detaching it from the instance and deleting it from the info cache [ 891.143652] env[63088]: DEBUG nova.network.neutron [req-04c9c65a-cbad-40ff-acf0-3ea530b80336 req-71b5b3fe-b289-4193-b3db-ceaf0983018f service nova] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 891.263850] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285100, 'name': CreateVM_Task, 'duration_secs': 1.669226} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.264041] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2964db35-7357-40a7-b4e6-7e2595549f5b] Created VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 891.264742] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.265359] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.265359] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 891.265492] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-97d6b1e3-8fa9-4eb1-b21b-3d334c2f7264 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.270270] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Waiting for the task: (returnval){ [ 891.270270] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]5212c9e1-1252-b5d7-1425-36116fbb1da7" [ 891.270270] env[63088]: _type = "Task" [ 891.270270] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.279524] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]5212c9e1-1252-b5d7-1425-36116fbb1da7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.339736] env[63088]: DEBUG nova.objects.base [None req-7e0b99c7-0e13-4f0c-9208-ec4af2ac98b6 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Object Instance<7b6aadb7-e34b-42b7-b69f-370434f5b665> lazy-loaded attributes: flavor,info_cache {{(pid=63088) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 891.410542] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Task: {'id': task-1285105, 'name': ReconfigVM_Task, 'duration_secs': 0.530662} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.411428] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: b50db882-598c-488f-b935-34f55f655642] Reconfigured VM instance instance-0000004a to attach disk [datastore1] b50db882-598c-488f-b935-34f55f655642/b50db882-598c-488f-b935-34f55f655642.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 891.411805] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9d0e5a68-6a7c-4400-80fa-5d1a976dcb62 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.418569] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Waiting for the task: (returnval){ [ 891.418569] env[63088]: value = "task-1285108" [ 891.418569] env[63088]: _type = "Task" [ 891.418569] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.420991] env[63088]: INFO nova.compute.claims [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 891.432584] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Task: {'id': task-1285108, 'name': Rename_Task} progress is 10%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.455492] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4ea5fd7b-e84b-41ba-9be2-088f6a9cab78 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Lock "8b564e29-96cf-4abf-963d-142b413fb464" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.342s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.481134] env[63088]: DEBUG nova.compute.manager [req-a5897d75-1684-4654-829c-6b7214b75f9f req-9e14b354-59aa-4ecd-b957-c601339b36fd service nova] [instance: 254db932-35f1-42e3-9207-cd886efd65a3] Received event network-vif-plugged-d2722fc9-4b54-495f-81bb-d5b411285079 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 891.481363] env[63088]: DEBUG oslo_concurrency.lockutils [req-a5897d75-1684-4654-829c-6b7214b75f9f req-9e14b354-59aa-4ecd-b957-c601339b36fd service nova] Acquiring lock "254db932-35f1-42e3-9207-cd886efd65a3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 891.481578] env[63088]: DEBUG oslo_concurrency.lockutils [req-a5897d75-1684-4654-829c-6b7214b75f9f req-9e14b354-59aa-4ecd-b957-c601339b36fd service nova] Lock "254db932-35f1-42e3-9207-cd886efd65a3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.481750] env[63088]: DEBUG oslo_concurrency.lockutils [req-a5897d75-1684-4654-829c-6b7214b75f9f req-9e14b354-59aa-4ecd-b957-c601339b36fd service nova] Lock "254db932-35f1-42e3-9207-cd886efd65a3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.481918] env[63088]: DEBUG nova.compute.manager [req-a5897d75-1684-4654-829c-6b7214b75f9f req-9e14b354-59aa-4ecd-b957-c601339b36fd service nova] [instance: 254db932-35f1-42e3-9207-cd886efd65a3] No waiting events found dispatching network-vif-plugged-d2722fc9-4b54-495f-81bb-d5b411285079 {{(pid=63088) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 891.482110] env[63088]: WARNING nova.compute.manager [req-a5897d75-1684-4654-829c-6b7214b75f9f req-9e14b354-59aa-4ecd-b957-c601339b36fd service nova] [instance: 254db932-35f1-42e3-9207-cd886efd65a3] Received unexpected event network-vif-plugged-d2722fc9-4b54-495f-81bb-d5b411285079 for instance with vm_state building and task_state spawning. [ 891.557045] env[63088]: DEBUG nova.network.neutron [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 254db932-35f1-42e3-9207-cd886efd65a3] Successfully updated port: d2722fc9-4b54-495f-81bb-d5b411285079 {{(pid=63088) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 891.568459] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Task: {'id': task-1285106, 'name': PowerOnVM_Task} progress is 90%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.571020] env[63088]: DEBUG oslo_vmware.api [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Task: {'id': task-1285107, 'name': PowerOffVM_Task, 'duration_secs': 0.227047} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.571938] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 891.572185] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Updating instance 'e6b0ce08-d67f-458f-92f5-1e904d03a5ef' progress to 17 {{(pid=63088) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 891.579272] env[63088]: DEBUG nova.network.neutron [-] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 891.649653] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5540ed77-eda8-44a3-a211-791fd53bec88 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.659352] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8521d769-227f-4600-8255-6372dd0a6dbf {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.693691] env[63088]: DEBUG nova.compute.manager [req-04c9c65a-cbad-40ff-acf0-3ea530b80336 req-71b5b3fe-b289-4193-b3db-ceaf0983018f service nova] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Detach interface failed, port_id=d5a10951-ae2e-48f5-b0ca-b1f144e5fe3a, reason: Instance 4971b24c-6710-4f50-9846-727dad264b1f could not be found. {{(pid=63088) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 891.783067] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]5212c9e1-1252-b5d7-1425-36116fbb1da7, 'name': SearchDatastore_Task, 'duration_secs': 0.011311} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.783067] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 891.783067] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: 2964db35-7357-40a7-b4e6-7e2595549f5b] Processing image 1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 891.783067] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.783352] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.783389] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 891.784027] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2aab53fe-9566-4f2a-a502-2c1a0bf4370c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.791836] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 891.791974] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 891.792915] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c1f2da51-e032-4886-8eda-2754d5f654c1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.798713] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Waiting for the task: (returnval){ [ 891.798713] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]5225f23c-bf78-c044-4113-39735792d1b1" [ 891.798713] env[63088]: _type = "Task" [ 891.798713] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.807645] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]5225f23c-bf78-c044-4113-39735792d1b1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.933246] env[63088]: INFO nova.compute.resource_tracker [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Updating resource usage from migration 5602a4d9-9018-483a-98f7-e1e412ff902a [ 891.935788] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Task: {'id': task-1285108, 'name': Rename_Task, 'duration_secs': 0.163323} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.936264] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: b50db882-598c-488f-b935-34f55f655642] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 891.936524] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-97f0e67d-6c77-4116-9bbd-fd1ba3df7263 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.943717] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Waiting for the task: (returnval){ [ 891.943717] env[63088]: value = "task-1285109" [ 891.943717] env[63088]: _type = "Task" [ 891.943717] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.951485] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Task: {'id': task-1285109, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.063585] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Acquiring lock "refresh_cache-254db932-35f1-42e3-9207-cd886efd65a3" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 892.063758] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Acquired lock "refresh_cache-254db932-35f1-42e3-9207-cd886efd65a3" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 892.063898] env[63088]: DEBUG nova.network.neutron [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 254db932-35f1-42e3-9207-cd886efd65a3] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 892.070302] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Task: {'id': task-1285106, 'name': PowerOnVM_Task, 'duration_secs': 1.093709} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.070977] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: 667fd5e9-5fe4-41e1-9d8a-896c6e5c6286] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 892.071298] env[63088]: INFO nova.compute.manager [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: 667fd5e9-5fe4-41e1-9d8a-896c6e5c6286] Took 10.62 seconds to spawn the instance on the hypervisor. [ 892.071517] env[63088]: DEBUG nova.compute.manager [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: 667fd5e9-5fe4-41e1-9d8a-896c6e5c6286] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 892.072285] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58082c38-2617-4e37-bffc-0cab0f3d218c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.078357] env[63088]: DEBUG nova.virt.hardware [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:33Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 892.078576] env[63088]: DEBUG nova.virt.hardware [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 892.078734] env[63088]: DEBUG nova.virt.hardware [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 892.078916] env[63088]: DEBUG nova.virt.hardware [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 892.079077] env[63088]: DEBUG nova.virt.hardware [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 892.079231] env[63088]: DEBUG nova.virt.hardware [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 892.079435] env[63088]: DEBUG nova.virt.hardware [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 892.079593] env[63088]: DEBUG nova.virt.hardware [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 892.079772] env[63088]: DEBUG nova.virt.hardware [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 892.080596] env[63088]: DEBUG nova.virt.hardware [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 892.080596] env[63088]: DEBUG nova.virt.hardware [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 892.089123] env[63088]: INFO nova.compute.manager [-] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Took 1.59 seconds to deallocate network for instance. [ 892.089352] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5f5ebce1-6544-42a3-bdc5-e23489a7cc78 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.103609] env[63088]: DEBUG nova.network.neutron [None req-7e0b99c7-0e13-4f0c-9208-ec4af2ac98b6 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Updating instance_info_cache with network_info: [{"id": "fa4a91d1-df9a-4789-bc5c-a8b95457cd93", "address": "fa:16:3e:7e:c2:c7", "network": {"id": "dfcbe153-bc01-4362-9247-de9f0b1c847b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-841757397-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.152", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dace8b5181b84623b08f903d12dfd31e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c68b7663-4f0e-47f0-ac7f-40c6d952f7bb", "external-id": "nsx-vlan-transportzone-696", "segmentation_id": 696, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfa4a91d1-df", "ovs_interfaceid": "fa4a91d1-df9a-4789-bc5c-a8b95457cd93", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 892.113052] env[63088]: DEBUG oslo_vmware.api [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Waiting for the task: (returnval){ [ 892.113052] env[63088]: value = "task-1285110" [ 892.113052] env[63088]: _type = "Task" [ 892.113052] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.125400] env[63088]: DEBUG oslo_vmware.api [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Task: {'id': task-1285110, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.226107] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c80fac3d-0e50-40aa-9ef8-8dd0ee502273 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.233259] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27f47ec9-4fb9-4f4a-8cb8-dcbdbf3f78e0 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.268385] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c18828f1-6121-4202-ac86-980b217ab2dd {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.276937] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe9236dc-6d70-4c00-8030-c63be02617e0 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.291650] env[63088]: DEBUG nova.compute.provider_tree [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 892.309240] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]5225f23c-bf78-c044-4113-39735792d1b1, 'name': SearchDatastore_Task, 'duration_secs': 0.008653} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.309990] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-598e0d3b-3b87-49f4-9736-8e48dadcacbc {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.316191] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Waiting for the task: (returnval){ [ 892.316191] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52b9774f-839c-eb07-11a8-814cfe26c50a" [ 892.316191] env[63088]: _type = "Task" [ 892.316191] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.324146] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52b9774f-839c-eb07-11a8-814cfe26c50a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.453967] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Task: {'id': task-1285109, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.598934] env[63088]: DEBUG nova.network.neutron [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 254db932-35f1-42e3-9207-cd886efd65a3] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 892.607560] env[63088]: DEBUG oslo_concurrency.lockutils [None req-7e0b99c7-0e13-4f0c-9208-ec4af2ac98b6 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Releasing lock "refresh_cache-7b6aadb7-e34b-42b7-b69f-370434f5b665" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 892.610112] env[63088]: INFO nova.compute.manager [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: 667fd5e9-5fe4-41e1-9d8a-896c6e5c6286] Took 33.95 seconds to build instance. [ 892.614712] env[63088]: DEBUG oslo_concurrency.lockutils [None req-125c7f05-a006-44cd-8334-62558a6d840d tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 892.627644] env[63088]: DEBUG oslo_vmware.api [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Task: {'id': task-1285110, 'name': ReconfigVM_Task, 'duration_secs': 0.217907} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.627969] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Updating instance 'e6b0ce08-d67f-458f-92f5-1e904d03a5ef' progress to 33 {{(pid=63088) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 892.738087] env[63088]: DEBUG nova.network.neutron [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 254db932-35f1-42e3-9207-cd886efd65a3] Updating instance_info_cache with network_info: [{"id": "d2722fc9-4b54-495f-81bb-d5b411285079", "address": "fa:16:3e:a8:64:fa", "network": {"id": "2bbaf351-fe63-46ba-ba7e-0f432c0a1bfb", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1438884226-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7bff1382b9694df08133c88a5fe783a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cca1f087-01e1-49ca-831b-5c51478a5d60", "external-id": "nsx-vlan-transportzone-439", "segmentation_id": 439, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd2722fc9-4b", "ovs_interfaceid": "d2722fc9-4b54-495f-81bb-d5b411285079", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 892.795298] env[63088]: DEBUG nova.scheduler.client.report [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 892.827736] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52b9774f-839c-eb07-11a8-814cfe26c50a, 'name': SearchDatastore_Task, 'duration_secs': 0.019575} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.827950] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 892.828236] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] 2964db35-7357-40a7-b4e6-7e2595549f5b/2964db35-7357-40a7-b4e6-7e2595549f5b.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 892.828487] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-10757577-8d59-4fa4-8dee-936018b16565 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.834119] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Waiting for the task: (returnval){ [ 892.834119] env[63088]: value = "task-1285111" [ 892.834119] env[63088]: _type = "Task" [ 892.834119] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.841358] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Task: {'id': task-1285111, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.953990] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Task: {'id': task-1285109, 'name': PowerOnVM_Task, 'duration_secs': 0.852576} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.954293] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: b50db882-598c-488f-b935-34f55f655642] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 892.954506] env[63088]: INFO nova.compute.manager [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: b50db882-598c-488f-b935-34f55f655642] Took 9.10 seconds to spawn the instance on the hypervisor. [ 892.954688] env[63088]: DEBUG nova.compute.manager [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: b50db882-598c-488f-b935-34f55f655642] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 892.955495] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a89656c9-502a-4384-93f3-8e4afdac5cd6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.111612] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e0b99c7-0e13-4f0c-9208-ec4af2ac98b6 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 893.112077] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Lock "667fd5e9-5fe4-41e1-9d8a-896c6e5c6286" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.464s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.112314] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b015f864-eb91-441c-be6c-0f7a6e74d60b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.120263] env[63088]: DEBUG oslo_vmware.api [None req-7e0b99c7-0e13-4f0c-9208-ec4af2ac98b6 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for the task: (returnval){ [ 893.120263] env[63088]: value = "task-1285112" [ 893.120263] env[63088]: _type = "Task" [ 893.120263] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.128941] env[63088]: DEBUG oslo_vmware.api [None req-7e0b99c7-0e13-4f0c-9208-ec4af2ac98b6 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285112, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.134527] env[63088]: DEBUG nova.virt.hardware [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:18:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='99d18326-0562-4fb9-afc4-e9e0d5c92bb4',id=38,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-1781853035',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 893.134880] env[63088]: DEBUG nova.virt.hardware [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 893.135095] env[63088]: DEBUG nova.virt.hardware [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 893.135322] env[63088]: DEBUG nova.virt.hardware [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 893.135545] env[63088]: DEBUG nova.virt.hardware [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 893.135772] env[63088]: DEBUG nova.virt.hardware [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 893.136042] env[63088]: DEBUG nova.virt.hardware [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 893.136252] env[63088]: DEBUG nova.virt.hardware [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 893.136511] env[63088]: DEBUG nova.virt.hardware [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 893.136760] env[63088]: DEBUG nova.virt.hardware [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 893.136948] env[63088]: DEBUG nova.virt.hardware [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 893.144139] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Reconfiguring VM instance instance-00000044 to detach disk 2000 {{(pid=63088) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 893.144972] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-17835515-579a-4ca0-90e5-6a76993287a7 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.166247] env[63088]: DEBUG oslo_vmware.api [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Waiting for the task: (returnval){ [ 893.166247] env[63088]: value = "task-1285113" [ 893.166247] env[63088]: _type = "Task" [ 893.166247] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.175255] env[63088]: DEBUG oslo_vmware.api [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Task: {'id': task-1285113, 'name': ReconfigVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.241314] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Releasing lock "refresh_cache-254db932-35f1-42e3-9207-cd886efd65a3" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 893.241778] env[63088]: DEBUG nova.compute.manager [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 254db932-35f1-42e3-9207-cd886efd65a3] Instance network_info: |[{"id": "d2722fc9-4b54-495f-81bb-d5b411285079", "address": "fa:16:3e:a8:64:fa", "network": {"id": "2bbaf351-fe63-46ba-ba7e-0f432c0a1bfb", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1438884226-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7bff1382b9694df08133c88a5fe783a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cca1f087-01e1-49ca-831b-5c51478a5d60", "external-id": "nsx-vlan-transportzone-439", "segmentation_id": 439, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd2722fc9-4b", "ovs_interfaceid": "d2722fc9-4b54-495f-81bb-d5b411285079", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 893.242344] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 254db932-35f1-42e3-9207-cd886efd65a3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a8:64:fa', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cca1f087-01e1-49ca-831b-5c51478a5d60', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd2722fc9-4b54-495f-81bb-d5b411285079', 'vif_model': 'vmxnet3'}] {{(pid=63088) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 893.251069] env[63088]: DEBUG oslo.service.loopingcall [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 893.251331] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 254db932-35f1-42e3-9207-cd886efd65a3] Creating VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 893.251589] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-23144237-1773-48f4-8bfb-f8843d5459d0 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.274232] env[63088]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 893.274232] env[63088]: value = "task-1285114" [ 893.274232] env[63088]: _type = "Task" [ 893.274232] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.286022] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285114, 'name': CreateVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.300206] env[63088]: DEBUG oslo_concurrency.lockutils [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.383s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.300458] env[63088]: INFO nova.compute.manager [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Migrating [ 893.307090] env[63088]: DEBUG oslo_concurrency.lockutils [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.089s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.308915] env[63088]: INFO nova.compute.claims [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 893.344626] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Task: {'id': task-1285111, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.481728} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.344917] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] 2964db35-7357-40a7-b4e6-7e2595549f5b/2964db35-7357-40a7-b4e6-7e2595549f5b.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 893.345254] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: 2964db35-7357-40a7-b4e6-7e2595549f5b] Extending root virtual disk to 1048576 {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 893.345478] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-077390ba-6dcf-4d5e-9b1d-8926f92f0dd5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.353429] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Waiting for the task: (returnval){ [ 893.353429] env[63088]: value = "task-1285115" [ 893.353429] env[63088]: _type = "Task" [ 893.353429] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.364897] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Task: {'id': task-1285115, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.474884] env[63088]: INFO nova.compute.manager [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: b50db882-598c-488f-b935-34f55f655642] Took 34.78 seconds to build instance. [ 893.528546] env[63088]: DEBUG nova.compute.manager [req-120e0e7d-d1da-4309-a16c-111a81bef085 req-bd152569-68f0-4f63-9cba-ac89b1a9f03a service nova] [instance: 254db932-35f1-42e3-9207-cd886efd65a3] Received event network-changed-d2722fc9-4b54-495f-81bb-d5b411285079 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 893.528778] env[63088]: DEBUG nova.compute.manager [req-120e0e7d-d1da-4309-a16c-111a81bef085 req-bd152569-68f0-4f63-9cba-ac89b1a9f03a service nova] [instance: 254db932-35f1-42e3-9207-cd886efd65a3] Refreshing instance network info cache due to event network-changed-d2722fc9-4b54-495f-81bb-d5b411285079. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 893.529624] env[63088]: DEBUG oslo_concurrency.lockutils [req-120e0e7d-d1da-4309-a16c-111a81bef085 req-bd152569-68f0-4f63-9cba-ac89b1a9f03a service nova] Acquiring lock "refresh_cache-254db932-35f1-42e3-9207-cd886efd65a3" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 893.529624] env[63088]: DEBUG oslo_concurrency.lockutils [req-120e0e7d-d1da-4309-a16c-111a81bef085 req-bd152569-68f0-4f63-9cba-ac89b1a9f03a service nova] Acquired lock "refresh_cache-254db932-35f1-42e3-9207-cd886efd65a3" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 893.529624] env[63088]: DEBUG nova.network.neutron [req-120e0e7d-d1da-4309-a16c-111a81bef085 req-bd152569-68f0-4f63-9cba-ac89b1a9f03a service nova] [instance: 254db932-35f1-42e3-9207-cd886efd65a3] Refreshing network info cache for port d2722fc9-4b54-495f-81bb-d5b411285079 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 893.630450] env[63088]: DEBUG oslo_vmware.api [None req-7e0b99c7-0e13-4f0c-9208-ec4af2ac98b6 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285112, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.676329] env[63088]: DEBUG oslo_vmware.api [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Task: {'id': task-1285113, 'name': ReconfigVM_Task, 'duration_secs': 0.286217} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.676838] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Reconfigured VM instance instance-00000044 to detach disk 2000 {{(pid=63088) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 893.677429] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9577f64-f130-44dd-ad39-af4aec0cce27 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.700611] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Reconfiguring VM instance instance-00000044 to attach disk [datastore2] e6b0ce08-d67f-458f-92f5-1e904d03a5ef/e6b0ce08-d67f-458f-92f5-1e904d03a5ef.vmdk or device None with type thin {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 893.700756] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0dc9f010-0ff6-441b-abee-10d212a09fad {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.718998] env[63088]: DEBUG oslo_vmware.api [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Waiting for the task: (returnval){ [ 893.718998] env[63088]: value = "task-1285116" [ 893.718998] env[63088]: _type = "Task" [ 893.718998] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.726807] env[63088]: DEBUG oslo_vmware.api [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Task: {'id': task-1285116, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.783988] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285114, 'name': CreateVM_Task, 'duration_secs': 0.438011} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.784204] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 254db932-35f1-42e3-9207-cd886efd65a3] Created VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 893.784879] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 893.785101] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 893.785463] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 893.785724] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c05a06e3-dc06-4e85-9a60-fc060d854880 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.789851] env[63088]: DEBUG oslo_vmware.api [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Waiting for the task: (returnval){ [ 893.789851] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]5260de76-a3ea-13c4-6e0d-0b3569e80f47" [ 893.789851] env[63088]: _type = "Task" [ 893.789851] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.797665] env[63088]: DEBUG oslo_vmware.api [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]5260de76-a3ea-13c4-6e0d-0b3569e80f47, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.821719] env[63088]: DEBUG oslo_concurrency.lockutils [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquiring lock "refresh_cache-2269e520-08b6-433f-8fe9-7b84d91e02d7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 893.821884] env[63088]: DEBUG oslo_concurrency.lockutils [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquired lock "refresh_cache-2269e520-08b6-433f-8fe9-7b84d91e02d7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 893.822085] env[63088]: DEBUG nova.network.neutron [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 893.861716] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Task: {'id': task-1285115, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066502} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.862011] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: 2964db35-7357-40a7-b4e6-7e2595549f5b] Extended root virtual disk {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 893.862788] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c041e2a-860a-4d71-a7de-dbae76e4aa7d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.884770] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: 2964db35-7357-40a7-b4e6-7e2595549f5b] Reconfiguring VM instance instance-0000004b to attach disk [datastore1] 2964db35-7357-40a7-b4e6-7e2595549f5b/2964db35-7357-40a7-b4e6-7e2595549f5b.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 893.884978] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1a336367-95ad-4bd4-8bfc-f6466316ba32 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.903849] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Waiting for the task: (returnval){ [ 893.903849] env[63088]: value = "task-1285117" [ 893.903849] env[63088]: _type = "Task" [ 893.903849] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.911437] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Task: {'id': task-1285117, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.976730] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Lock "b50db882-598c-488f-b935-34f55f655642" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 36.297s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.133132] env[63088]: DEBUG oslo_vmware.api [None req-7e0b99c7-0e13-4f0c-9208-ec4af2ac98b6 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285112, 'name': PowerOnVM_Task, 'duration_secs': 0.548137} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.135530] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e0b99c7-0e13-4f0c-9208-ec4af2ac98b6 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 894.135678] env[63088]: DEBUG nova.compute.manager [None req-7e0b99c7-0e13-4f0c-9208-ec4af2ac98b6 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 894.136456] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7acf0f4-37c4-4800-b503-e972b7a44404 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.229145] env[63088]: DEBUG oslo_vmware.api [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Task: {'id': task-1285116, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.257685] env[63088]: DEBUG nova.network.neutron [req-120e0e7d-d1da-4309-a16c-111a81bef085 req-bd152569-68f0-4f63-9cba-ac89b1a9f03a service nova] [instance: 254db932-35f1-42e3-9207-cd886efd65a3] Updated VIF entry in instance network info cache for port d2722fc9-4b54-495f-81bb-d5b411285079. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 894.258097] env[63088]: DEBUG nova.network.neutron [req-120e0e7d-d1da-4309-a16c-111a81bef085 req-bd152569-68f0-4f63-9cba-ac89b1a9f03a service nova] [instance: 254db932-35f1-42e3-9207-cd886efd65a3] Updating instance_info_cache with network_info: [{"id": "d2722fc9-4b54-495f-81bb-d5b411285079", "address": "fa:16:3e:a8:64:fa", "network": {"id": "2bbaf351-fe63-46ba-ba7e-0f432c0a1bfb", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1438884226-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7bff1382b9694df08133c88a5fe783a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cca1f087-01e1-49ca-831b-5c51478a5d60", "external-id": "nsx-vlan-transportzone-439", "segmentation_id": 439, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd2722fc9-4b", "ovs_interfaceid": "d2722fc9-4b54-495f-81bb-d5b411285079", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 894.300420] env[63088]: DEBUG oslo_vmware.api [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]5260de76-a3ea-13c4-6e0d-0b3569e80f47, 'name': SearchDatastore_Task, 'duration_secs': 0.013809} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.300722] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 894.300964] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 254db932-35f1-42e3-9207-cd886efd65a3] Processing image 1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 894.301249] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 894.301408] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 894.301596] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 894.301861] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1ce5c054-ed1b-489b-a520-ed4fe96e1df7 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.323330] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 894.323512] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 894.324230] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-09b40fa8-bbf3-4463-9200-6392714cf21f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.334017] env[63088]: DEBUG oslo_vmware.api [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Waiting for the task: (returnval){ [ 894.334017] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]5212447a-61a6-2dd4-21ba-12383b8e619f" [ 894.334017] env[63088]: _type = "Task" [ 894.334017] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.341707] env[63088]: DEBUG oslo_vmware.api [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]5212447a-61a6-2dd4-21ba-12383b8e619f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.416437] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Task: {'id': task-1285117, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.575343] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b6f538e-bafb-44f1-bea2-e9c9fd6e7818 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.582093] env[63088]: DEBUG nova.network.neutron [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Updating instance_info_cache with network_info: [{"id": "1fb0e170-03ab-4ee4-9f1f-c5c6008c823d", "address": "fa:16:3e:a8:02:36", "network": {"id": "dff14a7f-0af1-4e4e-a498-86d7c9816e6b", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-762073671-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f33f2701fad94864a8c406a404bc0a42", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "21310d90-efbc-45a8-a97f-c4358606530f", "external-id": "nsx-vlan-transportzone-672", "segmentation_id": 672, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1fb0e170-03", "ovs_interfaceid": "1fb0e170-03ab-4ee4-9f1f-c5c6008c823d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 894.591119] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-843b5202-a134-4753-accc-4a6624d23749 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.620753] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd589b71-fe9f-42e0-95b2-a7b87909ab7b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.628151] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70d4182e-8bfc-4f38-9265-f9e2d98f6c91 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.643770] env[63088]: DEBUG nova.compute.provider_tree [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 894.729592] env[63088]: DEBUG oslo_vmware.api [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Task: {'id': task-1285116, 'name': ReconfigVM_Task, 'duration_secs': 0.565325} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.729935] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Reconfigured VM instance instance-00000044 to attach disk [datastore2] e6b0ce08-d67f-458f-92f5-1e904d03a5ef/e6b0ce08-d67f-458f-92f5-1e904d03a5ef.vmdk or device None with type thin {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 894.730229] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Updating instance 'e6b0ce08-d67f-458f-92f5-1e904d03a5ef' progress to 50 {{(pid=63088) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 894.760381] env[63088]: DEBUG oslo_concurrency.lockutils [req-120e0e7d-d1da-4309-a16c-111a81bef085 req-bd152569-68f0-4f63-9cba-ac89b1a9f03a service nova] Releasing lock "refresh_cache-254db932-35f1-42e3-9207-cd886efd65a3" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 894.845760] env[63088]: DEBUG oslo_vmware.api [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]5212447a-61a6-2dd4-21ba-12383b8e619f, 'name': SearchDatastore_Task, 'duration_secs': 0.021214} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.846656] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-831aa187-562e-4f64-8653-e5bc42a7de12 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.851864] env[63088]: DEBUG oslo_vmware.api [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Waiting for the task: (returnval){ [ 894.851864] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52c271e4-b272-9cd5-f5d8-e8f9f98678bb" [ 894.851864] env[63088]: _type = "Task" [ 894.851864] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.859521] env[63088]: DEBUG oslo_vmware.api [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52c271e4-b272-9cd5-f5d8-e8f9f98678bb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.914304] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Task: {'id': task-1285117, 'name': ReconfigVM_Task, 'duration_secs': 0.783656} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.914597] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: 2964db35-7357-40a7-b4e6-7e2595549f5b] Reconfigured VM instance instance-0000004b to attach disk [datastore1] 2964db35-7357-40a7-b4e6-7e2595549f5b/2964db35-7357-40a7-b4e6-7e2595549f5b.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 894.915292] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2b06e398-3d10-479d-a578-36f03245846f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.923619] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Waiting for the task: (returnval){ [ 894.923619] env[63088]: value = "task-1285118" [ 894.923619] env[63088]: _type = "Task" [ 894.923619] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.933222] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Task: {'id': task-1285118, 'name': Rename_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.086490] env[63088]: DEBUG oslo_concurrency.lockutils [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Releasing lock "refresh_cache-2269e520-08b6-433f-8fe9-7b84d91e02d7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 895.148458] env[63088]: DEBUG nova.scheduler.client.report [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 895.238465] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24dca48c-9f91-42e5-bb78-44c5d818da77 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.257739] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-615e28d9-9f8a-4770-ad79-18f80dd8e774 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.275379] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Updating instance 'e6b0ce08-d67f-458f-92f5-1e904d03a5ef' progress to 67 {{(pid=63088) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 895.362272] env[63088]: DEBUG oslo_vmware.api [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52c271e4-b272-9cd5-f5d8-e8f9f98678bb, 'name': SearchDatastore_Task, 'duration_secs': 0.040954} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.362584] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 895.362863] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore2] 254db932-35f1-42e3-9207-cd886efd65a3/254db932-35f1-42e3-9207-cd886efd65a3.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 895.363174] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-86fe9669-6983-43da-acd2-ac8c3e1f75ad {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.369531] env[63088]: DEBUG oslo_vmware.api [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Waiting for the task: (returnval){ [ 895.369531] env[63088]: value = "task-1285119" [ 895.369531] env[63088]: _type = "Task" [ 895.369531] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.377634] env[63088]: DEBUG oslo_vmware.api [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285119, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.436279] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Task: {'id': task-1285118, 'name': Rename_Task, 'duration_secs': 0.284058} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.436582] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: 2964db35-7357-40a7-b4e6-7e2595549f5b] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 895.436877] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7be7fe65-c983-4f05-97c7-111c4f0dc4b4 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.444905] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Waiting for the task: (returnval){ [ 895.444905] env[63088]: value = "task-1285120" [ 895.444905] env[63088]: _type = "Task" [ 895.444905] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.453656] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Task: {'id': task-1285120, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.654910] env[63088]: DEBUG oslo_concurrency.lockutils [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.348s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.655489] env[63088]: DEBUG nova.compute.manager [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 895.658989] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0dea949d-9593-45db-b8f6-0d07193e67be tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.922s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 895.659259] env[63088]: DEBUG nova.objects.instance [None req-0dea949d-9593-45db-b8f6-0d07193e67be tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Lazy-loading 'resources' on Instance uuid 34ce411f-40c8-446e-b685-cd3ce07663dd {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 895.831035] env[63088]: DEBUG nova.network.neutron [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Port 63d8dbbe-4550-4d8d-8e85-9a9eadc81d82 binding to destination host cpu-1 is already ACTIVE {{(pid=63088) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 895.879738] env[63088]: DEBUG oslo_vmware.api [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285119, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.472907} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.880136] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore2] 254db932-35f1-42e3-9207-cd886efd65a3/254db932-35f1-42e3-9207-cd886efd65a3.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 895.880542] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 254db932-35f1-42e3-9207-cd886efd65a3] Extending root virtual disk to 1048576 {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 895.881022] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ac6fc323-a91c-42b6-ae1c-b69a20c908ee {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.888245] env[63088]: DEBUG oslo_vmware.api [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Waiting for the task: (returnval){ [ 895.888245] env[63088]: value = "task-1285121" [ 895.888245] env[63088]: _type = "Task" [ 895.888245] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.898166] env[63088]: DEBUG oslo_vmware.api [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285121, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.955436] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Task: {'id': task-1285120, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.165222] env[63088]: DEBUG nova.compute.utils [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 896.167095] env[63088]: DEBUG nova.compute.manager [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 896.167240] env[63088]: DEBUG nova.network.neutron [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 896.219557] env[63088]: DEBUG nova.policy [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd76b45ac31364495b86fad66b9acfb8a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2b18ddddf2314d83addf550b8cb91977', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 896.397775] env[63088]: DEBUG oslo_vmware.api [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285121, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.196539} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.398310] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 254db932-35f1-42e3-9207-cd886efd65a3] Extended root virtual disk {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 896.399331] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66b8ae12-9d32-4f51-aa28-8b41c2fc398c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.404657] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbf5c652-9ab2-452b-bd47-141fa2f2fb22 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.429796] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 254db932-35f1-42e3-9207-cd886efd65a3] Reconfiguring VM instance instance-0000004c to attach disk [datastore2] 254db932-35f1-42e3-9207-cd886efd65a3/254db932-35f1-42e3-9207-cd886efd65a3.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 896.430497] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7a88fcd7-f4e1-41df-8c26-cc243d033cc0 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.447433] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e465dced-cc7a-4e08-98b3-1408ef6deb86 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.455962] env[63088]: DEBUG oslo_vmware.api [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Waiting for the task: (returnval){ [ 896.455962] env[63088]: value = "task-1285122" [ 896.455962] env[63088]: _type = "Task" [ 896.455962] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.487174] env[63088]: DEBUG oslo_vmware.api [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Task: {'id': task-1285120, 'name': PowerOnVM_Task, 'duration_secs': 0.557223} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.488480] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38fed9cb-5b14-4eb8-aa87-518a59f95e23 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.491137] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: 2964db35-7357-40a7-b4e6-7e2595549f5b] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 896.491364] env[63088]: INFO nova.compute.manager [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: 2964db35-7357-40a7-b4e6-7e2595549f5b] Took 10.20 seconds to spawn the instance on the hypervisor. [ 896.491624] env[63088]: DEBUG nova.compute.manager [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: 2964db35-7357-40a7-b4e6-7e2595549f5b] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 896.495377] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbc224c2-34c2-4bff-aa44-16174a91c19e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.497933] env[63088]: DEBUG oslo_vmware.api [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285122, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.498725] env[63088]: DEBUG nova.network.neutron [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Successfully created port: 79ca8279-a3e2-4792-a4b5-57f899be5c41 {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 896.506687] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-203f13b3-fe1b-493c-ae01-142d5c7c3ef5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.522225] env[63088]: DEBUG nova.compute.provider_tree [None req-0dea949d-9593-45db-b8f6-0d07193e67be tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 896.603920] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1839c1e-b6df-4c07-94c1-497bb64258e2 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.623327] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Updating instance '2269e520-08b6-433f-8fe9-7b84d91e02d7' progress to 0 {{(pid=63088) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 896.670577] env[63088]: DEBUG nova.compute.manager [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 896.865285] env[63088]: DEBUG oslo_concurrency.lockutils [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Acquiring lock "e6b0ce08-d67f-458f-92f5-1e904d03a5ef-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 896.865670] env[63088]: DEBUG oslo_concurrency.lockutils [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Lock "e6b0ce08-d67f-458f-92f5-1e904d03a5ef-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.004s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.865950] env[63088]: DEBUG oslo_concurrency.lockutils [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Lock "e6b0ce08-d67f-458f-92f5-1e904d03a5ef-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.966143] env[63088]: DEBUG oslo_vmware.api [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285122, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.021570] env[63088]: INFO nova.compute.manager [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: 2964db35-7357-40a7-b4e6-7e2595549f5b] Took 38.30 seconds to build instance. [ 897.025025] env[63088]: DEBUG nova.scheduler.client.report [None req-0dea949d-9593-45db-b8f6-0d07193e67be tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 897.130642] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 897.131362] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c973c2d0-a2f8-4b5a-83d8-4c6a9be8d345 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.141549] env[63088]: DEBUG oslo_vmware.api [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Waiting for the task: (returnval){ [ 897.141549] env[63088]: value = "task-1285123" [ 897.141549] env[63088]: _type = "Task" [ 897.141549] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.149765] env[63088]: DEBUG oslo_vmware.api [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285123, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.466676] env[63088]: DEBUG oslo_vmware.api [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285122, 'name': ReconfigVM_Task, 'duration_secs': 0.809127} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.467030] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 254db932-35f1-42e3-9207-cd886efd65a3] Reconfigured VM instance instance-0000004c to attach disk [datastore2] 254db932-35f1-42e3-9207-cd886efd65a3/254db932-35f1-42e3-9207-cd886efd65a3.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 897.467675] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-23fae9ec-a3fa-4da0-99f2-e86e884a51d5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.473539] env[63088]: DEBUG oslo_vmware.api [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Waiting for the task: (returnval){ [ 897.473539] env[63088]: value = "task-1285124" [ 897.473539] env[63088]: _type = "Task" [ 897.473539] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.481210] env[63088]: DEBUG oslo_vmware.api [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285124, 'name': Rename_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.523671] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c764a25a-c70e-4aa8-9257-dbd6d542d193 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Lock "2964db35-7357-40a7-b4e6-7e2595549f5b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 39.815s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.529667] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0dea949d-9593-45db-b8f6-0d07193e67be tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.871s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.532160] env[63088]: DEBUG oslo_concurrency.lockutils [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.773s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 897.533696] env[63088]: INFO nova.compute.claims [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: bebc3318-24bb-4a37-8b23-66a12a7f7fd2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 897.553364] env[63088]: INFO nova.scheduler.client.report [None req-0dea949d-9593-45db-b8f6-0d07193e67be tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Deleted allocations for instance 34ce411f-40c8-446e-b685-cd3ce07663dd [ 897.652423] env[63088]: DEBUG oslo_vmware.api [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285123, 'name': PowerOffVM_Task} progress is 100%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.679367] env[63088]: DEBUG nova.compute.manager [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 897.706322] env[63088]: DEBUG nova.virt.hardware [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 897.706589] env[63088]: DEBUG nova.virt.hardware [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 897.707171] env[63088]: DEBUG nova.virt.hardware [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 897.707171] env[63088]: DEBUG nova.virt.hardware [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 897.707171] env[63088]: DEBUG nova.virt.hardware [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 897.707351] env[63088]: DEBUG nova.virt.hardware [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 897.707437] env[63088]: DEBUG nova.virt.hardware [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 897.707601] env[63088]: DEBUG nova.virt.hardware [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 897.707771] env[63088]: DEBUG nova.virt.hardware [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 897.708090] env[63088]: DEBUG nova.virt.hardware [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 897.708286] env[63088]: DEBUG nova.virt.hardware [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 897.709170] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7ad2541-c2d5-4f0a-9ad7-b73cb93ae7c5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.716894] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59b4a50e-c804-441f-9399-f0ba9a94ac1c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.785281] env[63088]: DEBUG oslo_concurrency.lockutils [None req-33fca862-7795-4260-b5ce-68be35c68550 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Acquiring lock "667fd5e9-5fe4-41e1-9d8a-896c6e5c6286" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 897.785586] env[63088]: DEBUG oslo_concurrency.lockutils [None req-33fca862-7795-4260-b5ce-68be35c68550 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Lock "667fd5e9-5fe4-41e1-9d8a-896c6e5c6286" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 897.785803] env[63088]: DEBUG oslo_concurrency.lockutils [None req-33fca862-7795-4260-b5ce-68be35c68550 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Acquiring lock "667fd5e9-5fe4-41e1-9d8a-896c6e5c6286-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 897.785991] env[63088]: DEBUG oslo_concurrency.lockutils [None req-33fca862-7795-4260-b5ce-68be35c68550 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Lock "667fd5e9-5fe4-41e1-9d8a-896c6e5c6286-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 897.786216] env[63088]: DEBUG oslo_concurrency.lockutils [None req-33fca862-7795-4260-b5ce-68be35c68550 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Lock "667fd5e9-5fe4-41e1-9d8a-896c6e5c6286-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.788402] env[63088]: INFO nova.compute.manager [None req-33fca862-7795-4260-b5ce-68be35c68550 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: 667fd5e9-5fe4-41e1-9d8a-896c6e5c6286] Terminating instance [ 897.790212] env[63088]: DEBUG nova.compute.manager [None req-33fca862-7795-4260-b5ce-68be35c68550 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: 667fd5e9-5fe4-41e1-9d8a-896c6e5c6286] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 897.790416] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-33fca862-7795-4260-b5ce-68be35c68550 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: 667fd5e9-5fe4-41e1-9d8a-896c6e5c6286] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 897.791354] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-684ebef2-b8fe-47d5-b2f5-18e953b297d8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.798458] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-33fca862-7795-4260-b5ce-68be35c68550 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: 667fd5e9-5fe4-41e1-9d8a-896c6e5c6286] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 897.798717] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-aadfd23c-a88b-49cd-8c62-57cfd12b5853 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.804802] env[63088]: DEBUG oslo_vmware.api [None req-33fca862-7795-4260-b5ce-68be35c68550 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Waiting for the task: (returnval){ [ 897.804802] env[63088]: value = "task-1285125" [ 897.804802] env[63088]: _type = "Task" [ 897.804802] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.813671] env[63088]: DEBUG oslo_vmware.api [None req-33fca862-7795-4260-b5ce-68be35c68550 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Task: {'id': task-1285125, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.892432] env[63088]: DEBUG nova.compute.manager [req-b7f2c5f1-d828-463a-8f29-2de5790da94b req-6a89b6bf-009a-4ab1-a977-38097d98047c service nova] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Received event network-vif-plugged-79ca8279-a3e2-4792-a4b5-57f899be5c41 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 897.892666] env[63088]: DEBUG oslo_concurrency.lockutils [req-b7f2c5f1-d828-463a-8f29-2de5790da94b req-6a89b6bf-009a-4ab1-a977-38097d98047c service nova] Acquiring lock "71e37d8e-a454-46c4-a3cc-3d5671a32beb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 897.892886] env[63088]: DEBUG oslo_concurrency.lockutils [req-b7f2c5f1-d828-463a-8f29-2de5790da94b req-6a89b6bf-009a-4ab1-a977-38097d98047c service nova] Lock "71e37d8e-a454-46c4-a3cc-3d5671a32beb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 897.893070] env[63088]: DEBUG oslo_concurrency.lockutils [req-b7f2c5f1-d828-463a-8f29-2de5790da94b req-6a89b6bf-009a-4ab1-a977-38097d98047c service nova] Lock "71e37d8e-a454-46c4-a3cc-3d5671a32beb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.893255] env[63088]: DEBUG nova.compute.manager [req-b7f2c5f1-d828-463a-8f29-2de5790da94b req-6a89b6bf-009a-4ab1-a977-38097d98047c service nova] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] No waiting events found dispatching network-vif-plugged-79ca8279-a3e2-4792-a4b5-57f899be5c41 {{(pid=63088) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 897.893441] env[63088]: WARNING nova.compute.manager [req-b7f2c5f1-d828-463a-8f29-2de5790da94b req-6a89b6bf-009a-4ab1-a977-38097d98047c service nova] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Received unexpected event network-vif-plugged-79ca8279-a3e2-4792-a4b5-57f899be5c41 for instance with vm_state building and task_state spawning. [ 897.949571] env[63088]: DEBUG oslo_concurrency.lockutils [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Acquiring lock "refresh_cache-e6b0ce08-d67f-458f-92f5-1e904d03a5ef" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 897.949827] env[63088]: DEBUG oslo_concurrency.lockutils [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Acquired lock "refresh_cache-e6b0ce08-d67f-458f-92f5-1e904d03a5ef" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 897.950101] env[63088]: DEBUG nova.network.neutron [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 897.983184] env[63088]: DEBUG oslo_vmware.api [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285124, 'name': Rename_Task, 'duration_secs': 0.179752} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.983998] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 254db932-35f1-42e3-9207-cd886efd65a3] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 897.983998] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ef4bec24-15fb-4f8a-976b-1e62042f57e1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.989340] env[63088]: DEBUG oslo_vmware.api [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Waiting for the task: (returnval){ [ 897.989340] env[63088]: value = "task-1285126" [ 897.989340] env[63088]: _type = "Task" [ 897.989340] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.997819] env[63088]: DEBUG oslo_vmware.api [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285126, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.003434] env[63088]: DEBUG nova.network.neutron [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Successfully updated port: 79ca8279-a3e2-4792-a4b5-57f899be5c41 {{(pid=63088) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 898.061854] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0dea949d-9593-45db-b8f6-0d07193e67be tempest-ServersV294TestFqdnHostnames-799621345 tempest-ServersV294TestFqdnHostnames-799621345-project-member] Lock "34ce411f-40c8-446e-b685-cd3ce07663dd" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.701s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.156490] env[63088]: DEBUG oslo_vmware.api [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285123, 'name': PowerOffVM_Task, 'duration_secs': 0.548727} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.156830] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 898.156959] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Updating instance '2269e520-08b6-433f-8fe9-7b84d91e02d7' progress to 17 {{(pid=63088) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 898.317879] env[63088]: DEBUG oslo_vmware.api [None req-33fca862-7795-4260-b5ce-68be35c68550 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Task: {'id': task-1285125, 'name': PowerOffVM_Task, 'duration_secs': 0.27979} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.317879] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-33fca862-7795-4260-b5ce-68be35c68550 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: 667fd5e9-5fe4-41e1-9d8a-896c6e5c6286] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 898.317879] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-33fca862-7795-4260-b5ce-68be35c68550 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: 667fd5e9-5fe4-41e1-9d8a-896c6e5c6286] Unregistering the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 898.317879] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-35a5ad48-dcb2-4494-873e-5a23dbb28630 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.384025] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-33fca862-7795-4260-b5ce-68be35c68550 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: 667fd5e9-5fe4-41e1-9d8a-896c6e5c6286] Unregistered the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 898.384025] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-33fca862-7795-4260-b5ce-68be35c68550 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: 667fd5e9-5fe4-41e1-9d8a-896c6e5c6286] Deleting contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 898.384258] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-33fca862-7795-4260-b5ce-68be35c68550 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Deleting the datastore file [datastore1] 667fd5e9-5fe4-41e1-9d8a-896c6e5c6286 {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 898.384431] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5d2f9d29-abdc-402d-85c5-24707f65b39d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.393431] env[63088]: DEBUG oslo_vmware.api [None req-33fca862-7795-4260-b5ce-68be35c68550 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Waiting for the task: (returnval){ [ 898.393431] env[63088]: value = "task-1285128" [ 898.393431] env[63088]: _type = "Task" [ 898.393431] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.401342] env[63088]: DEBUG oslo_vmware.api [None req-33fca862-7795-4260-b5ce-68be35c68550 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Task: {'id': task-1285128, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.502965] env[63088]: DEBUG oslo_vmware.api [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285126, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.505708] env[63088]: DEBUG oslo_concurrency.lockutils [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Acquiring lock "refresh_cache-71e37d8e-a454-46c4-a3cc-3d5671a32beb" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 898.505708] env[63088]: DEBUG oslo_concurrency.lockutils [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Acquired lock "refresh_cache-71e37d8e-a454-46c4-a3cc-3d5671a32beb" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.505852] env[63088]: DEBUG nova.network.neutron [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 898.663985] env[63088]: DEBUG nova.virt.hardware [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:33Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 898.664273] env[63088]: DEBUG nova.virt.hardware [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 898.664438] env[63088]: DEBUG nova.virt.hardware [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 898.664624] env[63088]: DEBUG nova.virt.hardware [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 898.664776] env[63088]: DEBUG nova.virt.hardware [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 898.665090] env[63088]: DEBUG nova.virt.hardware [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 898.669017] env[63088]: DEBUG nova.virt.hardware [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 898.669017] env[63088]: DEBUG nova.virt.hardware [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 898.669017] env[63088]: DEBUG nova.virt.hardware [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 898.669017] env[63088]: DEBUG nova.virt.hardware [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 898.669017] env[63088]: DEBUG nova.virt.hardware [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 898.673888] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-89967fff-bbac-4140-ada9-8e202a87836e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.693882] env[63088]: DEBUG oslo_vmware.api [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Waiting for the task: (returnval){ [ 898.693882] env[63088]: value = "task-1285129" [ 898.693882] env[63088]: _type = "Task" [ 898.693882] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.707165] env[63088]: DEBUG oslo_vmware.api [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285129, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.840244] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe2e7bc1-b0ae-48b5-bd69-afba67d9d6e9 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.852015] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82cdf52c-0a1f-4cbb-b716-f4542c8d01a0 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.881648] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53c05c62-0f2e-4a98-ab78-0647ca2b620b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.889012] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf449361-e20a-4462-94f8-698aa49a366c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.904888] env[63088]: DEBUG nova.compute.provider_tree [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 898.911086] env[63088]: DEBUG oslo_vmware.api [None req-33fca862-7795-4260-b5ce-68be35c68550 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Task: {'id': task-1285128, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.180195} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.911394] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-33fca862-7795-4260-b5ce-68be35c68550 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 898.911576] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-33fca862-7795-4260-b5ce-68be35c68550 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: 667fd5e9-5fe4-41e1-9d8a-896c6e5c6286] Deleted contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 898.911760] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-33fca862-7795-4260-b5ce-68be35c68550 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: 667fd5e9-5fe4-41e1-9d8a-896c6e5c6286] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 898.911937] env[63088]: INFO nova.compute.manager [None req-33fca862-7795-4260-b5ce-68be35c68550 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: 667fd5e9-5fe4-41e1-9d8a-896c6e5c6286] Took 1.12 seconds to destroy the instance on the hypervisor. [ 898.912206] env[63088]: DEBUG oslo.service.loopingcall [None req-33fca862-7795-4260-b5ce-68be35c68550 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 898.914864] env[63088]: DEBUG nova.compute.manager [-] [instance: 667fd5e9-5fe4-41e1-9d8a-896c6e5c6286] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 898.914985] env[63088]: DEBUG nova.network.neutron [-] [instance: 667fd5e9-5fe4-41e1-9d8a-896c6e5c6286] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 899.001689] env[63088]: DEBUG oslo_vmware.api [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285126, 'name': PowerOnVM_Task, 'duration_secs': 0.638676} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.001968] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 254db932-35f1-42e3-9207-cd886efd65a3] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 899.003459] env[63088]: INFO nova.compute.manager [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 254db932-35f1-42e3-9207-cd886efd65a3] Took 7.95 seconds to spawn the instance on the hypervisor. [ 899.003459] env[63088]: DEBUG nova.compute.manager [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 254db932-35f1-42e3-9207-cd886efd65a3] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 899.007126] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94896d61-b368-4e26-9e8a-b6a41ee19076 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.061672] env[63088]: DEBUG nova.network.neutron [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Updating instance_info_cache with network_info: [{"id": "63d8dbbe-4550-4d8d-8e85-9a9eadc81d82", "address": "fa:16:3e:d1:72:56", "network": {"id": "ebeefbc2-043c-4e76-aeaf-4f9113869855", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.60", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "bfb6bc5062fc4d51af1d3c577659be9a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271fe7a0-dfd7-409b-920a-cf83ef1a86a3", "external-id": "nsx-vlan-transportzone-728", "segmentation_id": 728, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap63d8dbbe-45", "ovs_interfaceid": "63d8dbbe-4550-4d8d-8e85-9a9eadc81d82", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 899.084699] env[63088]: DEBUG nova.network.neutron [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 899.204801] env[63088]: DEBUG oslo_vmware.api [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285129, 'name': ReconfigVM_Task, 'duration_secs': 0.163728} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.205346] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Updating instance '2269e520-08b6-433f-8fe9-7b84d91e02d7' progress to 33 {{(pid=63088) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 899.263393] env[63088]: DEBUG nova.compute.manager [req-770199bb-7429-474e-a65c-f1625936d1ee req-def51768-dcb4-42cb-b66d-be3a00537592 service nova] [instance: 667fd5e9-5fe4-41e1-9d8a-896c6e5c6286] Received event network-vif-deleted-602aeee4-c8d8-41c6-b145-dae2eadfc9a7 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 899.263393] env[63088]: INFO nova.compute.manager [req-770199bb-7429-474e-a65c-f1625936d1ee req-def51768-dcb4-42cb-b66d-be3a00537592 service nova] [instance: 667fd5e9-5fe4-41e1-9d8a-896c6e5c6286] Neutron deleted interface 602aeee4-c8d8-41c6-b145-dae2eadfc9a7; detaching it from the instance and deleting it from the info cache [ 899.263393] env[63088]: DEBUG nova.network.neutron [req-770199bb-7429-474e-a65c-f1625936d1ee req-def51768-dcb4-42cb-b66d-be3a00537592 service nova] [instance: 667fd5e9-5fe4-41e1-9d8a-896c6e5c6286] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 899.287516] env[63088]: DEBUG nova.network.neutron [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Updating instance_info_cache with network_info: [{"id": "79ca8279-a3e2-4792-a4b5-57f899be5c41", "address": "fa:16:3e:53:f8:3c", "network": {"id": "cc2c0304-8906-443a-a76e-e7a32158346f", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1867758042-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b18ddddf2314d83addf550b8cb91977", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0f096917-a0cf-4add-a9d2-23ca1c723b3b", "external-id": "nsx-vlan-transportzone-894", "segmentation_id": 894, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap79ca8279-a3", "ovs_interfaceid": "79ca8279-a3e2-4792-a4b5-57f899be5c41", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 899.320112] env[63088]: DEBUG oslo_concurrency.lockutils [None req-e2dee0e1-16ef-4ce0-9a3b-bc5d82868451 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Acquiring lock "dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.320112] env[63088]: DEBUG oslo_concurrency.lockutils [None req-e2dee0e1-16ef-4ce0-9a3b-bc5d82868451 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Lock "dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.408015] env[63088]: DEBUG nova.scheduler.client.report [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 899.522364] env[63088]: INFO nova.compute.manager [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 254db932-35f1-42e3-9207-cd886efd65a3] Took 27.81 seconds to build instance. [ 899.564252] env[63088]: DEBUG oslo_concurrency.lockutils [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Releasing lock "refresh_cache-e6b0ce08-d67f-458f-92f5-1e904d03a5ef" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 899.712919] env[63088]: DEBUG nova.virt.hardware [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 899.713200] env[63088]: DEBUG nova.virt.hardware [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 899.713372] env[63088]: DEBUG nova.virt.hardware [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 899.713875] env[63088]: DEBUG nova.virt.hardware [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 899.713875] env[63088]: DEBUG nova.virt.hardware [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 899.713875] env[63088]: DEBUG nova.virt.hardware [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 899.714190] env[63088]: DEBUG nova.virt.hardware [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 899.714355] env[63088]: DEBUG nova.virt.hardware [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 899.714549] env[63088]: DEBUG nova.virt.hardware [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 899.714939] env[63088]: DEBUG nova.virt.hardware [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 899.715240] env[63088]: DEBUG nova.virt.hardware [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 899.720551] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Reconfiguring VM instance instance-00000047 to detach disk 2000 {{(pid=63088) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 899.720862] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-78d20c21-5cd5-4510-9716-31b58a71524f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.734294] env[63088]: DEBUG nova.network.neutron [-] [instance: 667fd5e9-5fe4-41e1-9d8a-896c6e5c6286] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 899.741598] env[63088]: DEBUG oslo_vmware.api [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Waiting for the task: (returnval){ [ 899.741598] env[63088]: value = "task-1285130" [ 899.741598] env[63088]: _type = "Task" [ 899.741598] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.750457] env[63088]: DEBUG oslo_vmware.api [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285130, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.765510] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-940f2b8a-f79e-4e41-b91e-2ad3be89399c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.776564] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fd9a908-5b50-433d-ac74-2205cb8b836f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.788980] env[63088]: DEBUG oslo_concurrency.lockutils [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Releasing lock "refresh_cache-71e37d8e-a454-46c4-a3cc-3d5671a32beb" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 899.789328] env[63088]: DEBUG nova.compute.manager [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Instance network_info: |[{"id": "79ca8279-a3e2-4792-a4b5-57f899be5c41", "address": "fa:16:3e:53:f8:3c", "network": {"id": "cc2c0304-8906-443a-a76e-e7a32158346f", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1867758042-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b18ddddf2314d83addf550b8cb91977", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0f096917-a0cf-4add-a9d2-23ca1c723b3b", "external-id": "nsx-vlan-transportzone-894", "segmentation_id": 894, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap79ca8279-a3", "ovs_interfaceid": "79ca8279-a3e2-4792-a4b5-57f899be5c41", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 899.789786] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:53:f8:3c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0f096917-a0cf-4add-a9d2-23ca1c723b3b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '79ca8279-a3e2-4792-a4b5-57f899be5c41', 'vif_model': 'vmxnet3'}] {{(pid=63088) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 899.797639] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Creating folder: Project (2b18ddddf2314d83addf550b8cb91977). Parent ref: group-v275816. {{(pid=63088) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 900.551806] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bd93052b-dbb2-4bd4-87e6-b60498273b7c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.555193] env[63088]: DEBUG nova.compute.utils [None req-e2dee0e1-16ef-4ce0-9a3b-bc5d82868451 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 900.558541] env[63088]: DEBUG oslo_concurrency.lockutils [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.026s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 900.559031] env[63088]: DEBUG nova.compute.manager [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: bebc3318-24bb-4a37-8b23-66a12a7f7fd2] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 900.561727] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a202c6b4-90b7-4d85-aaed-a77f81165221 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Lock "254db932-35f1-42e3-9207-cd886efd65a3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.861s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 900.565031] env[63088]: INFO nova.compute.manager [-] [instance: 667fd5e9-5fe4-41e1-9d8a-896c6e5c6286] Took 1.65 seconds to deallocate network for instance. [ 900.565455] env[63088]: DEBUG nova.compute.manager [req-770199bb-7429-474e-a65c-f1625936d1ee req-def51768-dcb4-42cb-b66d-be3a00537592 service nova] [instance: 667fd5e9-5fe4-41e1-9d8a-896c6e5c6286] Detach interface failed, port_id=602aeee4-c8d8-41c6-b145-dae2eadfc9a7, reason: Instance 667fd5e9-5fe4-41e1-9d8a-896c6e5c6286 could not be found. {{(pid=63088) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 900.572110] env[63088]: DEBUG nova.compute.manager [req-f82b1e91-bf6e-4570-857a-f6c127fc907f req-dfcdb642-57c7-451f-a946-245ade281ede service nova] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Received event network-changed-79ca8279-a3e2-4792-a4b5-57f899be5c41 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 900.572110] env[63088]: DEBUG nova.compute.manager [req-f82b1e91-bf6e-4570-857a-f6c127fc907f req-dfcdb642-57c7-451f-a946-245ade281ede service nova] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Refreshing instance network info cache due to event network-changed-79ca8279-a3e2-4792-a4b5-57f899be5c41. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 900.572110] env[63088]: DEBUG oslo_concurrency.lockutils [req-f82b1e91-bf6e-4570-857a-f6c127fc907f req-dfcdb642-57c7-451f-a946-245ade281ede service nova] Acquiring lock "refresh_cache-71e37d8e-a454-46c4-a3cc-3d5671a32beb" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 900.572110] env[63088]: DEBUG oslo_concurrency.lockutils [req-f82b1e91-bf6e-4570-857a-f6c127fc907f req-dfcdb642-57c7-451f-a946-245ade281ede service nova] Acquired lock "refresh_cache-71e37d8e-a454-46c4-a3cc-3d5671a32beb" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 900.572110] env[63088]: DEBUG nova.network.neutron [req-f82b1e91-bf6e-4570-857a-f6c127fc907f req-dfcdb642-57c7-451f-a946-245ade281ede service nova] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Refreshing network info cache for port 79ca8279-a3e2-4792-a4b5-57f899be5c41 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 900.572697] env[63088]: DEBUG oslo_concurrency.lockutils [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.772s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 900.574370] env[63088]: INFO nova.compute.claims [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 45513c6b-c000-4ee1-8893-4e084ffc22c5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 900.592131] env[63088]: DEBUG oslo_vmware.api [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285130, 'name': ReconfigVM_Task, 'duration_secs': 0.262572} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.592935] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Reconfigured VM instance instance-00000047 to detach disk 2000 {{(pid=63088) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 900.594441] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb306029-79e8-447e-8c00-396b6271e417 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.598844] env[63088]: INFO nova.virt.vmwareapi.vm_util [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Created folder: Project (2b18ddddf2314d83addf550b8cb91977) in parent group-v275816. [ 900.598844] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Creating folder: Instances. Parent ref: group-v275898. {{(pid=63088) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 900.598971] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ebc07e40-7e1e-468c-9987-504ca0b51607 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.619126] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Reconfiguring VM instance instance-00000047 to attach disk [datastore2] 2269e520-08b6-433f-8fe9-7b84d91e02d7/2269e520-08b6-433f-8fe9-7b84d91e02d7.vmdk or device None with type thin {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 900.622418] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c551bd6d-4c90-439d-a8df-71d0b4803f07 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.636492] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58b67bcd-82d5-411f-8d47-e208fbbf0ea4 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.640553] env[63088]: INFO nova.virt.vmwareapi.vm_util [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Created folder: Instances in parent group-v275898. [ 900.640792] env[63088]: DEBUG oslo.service.loopingcall [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 900.641706] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Creating VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 900.641949] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fee2f0d4-c6a4-4229-b740-ed9be5765935 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.670724] env[63088]: DEBUG oslo_vmware.api [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Waiting for the task: (returnval){ [ 900.670724] env[63088]: value = "task-1285133" [ 900.670724] env[63088]: _type = "Task" [ 900.670724] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.672351] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29d5fd4c-42a2-4cd8-b181-ae79c4166f90 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.679363] env[63088]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 900.679363] env[63088]: value = "task-1285134" [ 900.679363] env[63088]: _type = "Task" [ 900.679363] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.688330] env[63088]: DEBUG oslo_vmware.api [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285133, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.688647] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Updating instance 'e6b0ce08-d67f-458f-92f5-1e904d03a5ef' progress to 83 {{(pid=63088) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 900.696601] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285134, 'name': CreateVM_Task} progress is 6%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.069577] env[63088]: DEBUG oslo_concurrency.lockutils [None req-e2dee0e1-16ef-4ce0-9a3b-bc5d82868451 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Lock "dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.750s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.071096] env[63088]: DEBUG nova.compute.utils [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 901.073049] env[63088]: DEBUG nova.compute.manager [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: bebc3318-24bb-4a37-8b23-66a12a7f7fd2] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 901.073333] env[63088]: DEBUG nova.network.neutron [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: bebc3318-24bb-4a37-8b23-66a12a7f7fd2] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 901.089363] env[63088]: DEBUG oslo_concurrency.lockutils [None req-33fca862-7795-4260-b5ce-68be35c68550 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.155966] env[63088]: DEBUG nova.policy [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '37d241cdf58e4aa7b05f1f24a3c78296', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e27be174c0b3400193f9ae64a242da1a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 901.181202] env[63088]: DEBUG oslo_vmware.api [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285133, 'name': ReconfigVM_Task, 'duration_secs': 0.309763} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.184303] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Reconfigured VM instance instance-00000047 to attach disk [datastore2] 2269e520-08b6-433f-8fe9-7b84d91e02d7/2269e520-08b6-433f-8fe9-7b84d91e02d7.vmdk or device None with type thin {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 901.184590] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Updating instance '2269e520-08b6-433f-8fe9-7b84d91e02d7' progress to 50 {{(pid=63088) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 901.201104] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 901.201460] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285134, 'name': CreateVM_Task, 'duration_secs': 0.304286} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.201721] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a3483d03-5864-4b76-8f39-8118050d286a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.209157] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Created VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 901.209157] env[63088]: DEBUG oslo_concurrency.lockutils [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 901.209157] env[63088]: DEBUG oslo_concurrency.lockutils [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.209342] env[63088]: DEBUG oslo_concurrency.lockutils [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 901.210094] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e101d6ca-442e-482a-b483-3637c68dad20 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.216297] env[63088]: DEBUG oslo_vmware.api [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Waiting for the task: (returnval){ [ 901.216297] env[63088]: value = "task-1285135" [ 901.216297] env[63088]: _type = "Task" [ 901.216297] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.223037] env[63088]: DEBUG oslo_vmware.api [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Waiting for the task: (returnval){ [ 901.223037] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52c13c8a-8379-8f0b-91b3-0c9baa37ae89" [ 901.223037] env[63088]: _type = "Task" [ 901.223037] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.229498] env[63088]: DEBUG oslo_vmware.api [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Task: {'id': task-1285135, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.235020] env[63088]: DEBUG oslo_vmware.api [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52c13c8a-8379-8f0b-91b3-0c9baa37ae89, 'name': SearchDatastore_Task, 'duration_secs': 0.012306} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.235370] env[63088]: DEBUG oslo_concurrency.lockutils [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 901.235635] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Processing image 1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 901.235872] env[63088]: DEBUG oslo_concurrency.lockutils [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 901.236026] env[63088]: DEBUG oslo_concurrency.lockutils [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.236210] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 901.236473] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ba2c828a-2d0f-4d05-ba4e-3fc0ddf34e1b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.245091] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 901.245289] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 901.245979] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-75a730aa-7c74-4813-8744-c5a3c72def5e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.252049] env[63088]: DEBUG oslo_vmware.api [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Waiting for the task: (returnval){ [ 901.252049] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52b83dfe-bd9d-54a2-3884-3b949a498c8e" [ 901.252049] env[63088]: _type = "Task" [ 901.252049] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.261467] env[63088]: DEBUG oslo_vmware.api [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52b83dfe-bd9d-54a2-3884-3b949a498c8e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.509383] env[63088]: DEBUG nova.network.neutron [req-f82b1e91-bf6e-4570-857a-f6c127fc907f req-dfcdb642-57c7-451f-a946-245ade281ede service nova] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Updated VIF entry in instance network info cache for port 79ca8279-a3e2-4792-a4b5-57f899be5c41. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 901.509797] env[63088]: DEBUG nova.network.neutron [req-f82b1e91-bf6e-4570-857a-f6c127fc907f req-dfcdb642-57c7-451f-a946-245ade281ede service nova] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Updating instance_info_cache with network_info: [{"id": "79ca8279-a3e2-4792-a4b5-57f899be5c41", "address": "fa:16:3e:53:f8:3c", "network": {"id": "cc2c0304-8906-443a-a76e-e7a32158346f", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1867758042-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b18ddddf2314d83addf550b8cb91977", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0f096917-a0cf-4add-a9d2-23ca1c723b3b", "external-id": "nsx-vlan-transportzone-894", "segmentation_id": 894, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap79ca8279-a3", "ovs_interfaceid": "79ca8279-a3e2-4792-a4b5-57f899be5c41", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 901.577580] env[63088]: DEBUG nova.compute.manager [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: bebc3318-24bb-4a37-8b23-66a12a7f7fd2] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 901.594759] env[63088]: DEBUG nova.compute.manager [None req-b81753c4-9726-4ca0-bfa8-881e9fef1e6f tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 254db932-35f1-42e3-9207-cd886efd65a3] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 901.596500] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62694891-996b-44e5-ba82-f6635b922007 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.653360] env[63088]: DEBUG nova.network.neutron [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: bebc3318-24bb-4a37-8b23-66a12a7f7fd2] Successfully created port: bd817fe9-5d27-40bf-9348-e29243b263f5 {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 901.695916] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8ea9e4f-8ab7-4c1e-9ec5-90fe16e7e679 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.714345] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8a704bf-1f9c-44ad-8e11-6b3991c4d566 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.734872] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Updating instance '2269e520-08b6-433f-8fe9-7b84d91e02d7' progress to 67 {{(pid=63088) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 901.745732] env[63088]: DEBUG oslo_vmware.api [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Task: {'id': task-1285135, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.764082] env[63088]: DEBUG oslo_vmware.api [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52b83dfe-bd9d-54a2-3884-3b949a498c8e, 'name': SearchDatastore_Task, 'duration_secs': 0.008336} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.764989] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aff98056-9b50-47ea-9638-90ec85f7d108 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.774104] env[63088]: DEBUG oslo_vmware.api [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Waiting for the task: (returnval){ [ 901.774104] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52ce2e97-6cf5-51e3-7a17-5dc10a65a3d6" [ 901.774104] env[63088]: _type = "Task" [ 901.774104] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.782113] env[63088]: DEBUG oslo_vmware.api [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52ce2e97-6cf5-51e3-7a17-5dc10a65a3d6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.904285] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49442b03-3353-41e5-9718-c9f7ad3667ab {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.910654] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f7fbb3c-f374-4e2d-bf21-b875f1028152 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.946335] env[63088]: DEBUG oslo_concurrency.lockutils [None req-e2dee0e1-16ef-4ce0-9a3b-bc5d82868451 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Acquiring lock "dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.946335] env[63088]: DEBUG oslo_concurrency.lockutils [None req-e2dee0e1-16ef-4ce0-9a3b-bc5d82868451 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Lock "dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.946335] env[63088]: INFO nova.compute.manager [None req-e2dee0e1-16ef-4ce0-9a3b-bc5d82868451 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Attaching volume deedf893-eb6c-4ff0-ab0b-3e34abe88c8f to /dev/sdb [ 901.947726] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b0559f8-a551-4a9d-88c8-d3f21bb8b736 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.957654] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83ac7d06-ee4a-4ffc-b4ee-6a8534b5acfd {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.971779] env[63088]: DEBUG nova.compute.provider_tree [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 901.985989] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff9eb3f8-2e07-4984-97df-45fe5ea20500 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.993270] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-112c69e5-b3ae-4f3e-81c2-1cff41073628 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.006782] env[63088]: DEBUG nova.virt.block_device [None req-e2dee0e1-16ef-4ce0-9a3b-bc5d82868451 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Updating existing volume attachment record: f7df46eb-8b12-43d5-8972-9b6a334395f0 {{(pid=63088) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 902.012578] env[63088]: DEBUG oslo_concurrency.lockutils [req-f82b1e91-bf6e-4570-857a-f6c127fc907f req-dfcdb642-57c7-451f-a946-245ade281ede service nova] Releasing lock "refresh_cache-71e37d8e-a454-46c4-a3cc-3d5671a32beb" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 902.111775] env[63088]: INFO nova.compute.manager [None req-b81753c4-9726-4ca0-bfa8-881e9fef1e6f tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 254db932-35f1-42e3-9207-cd886efd65a3] instance snapshotting [ 902.114714] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa339ab9-8622-4403-8121-f73258849330 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.134118] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6272a5d4-f6e2-41bf-9202-80ef067bcc71 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.226434] env[63088]: DEBUG oslo_vmware.api [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Task: {'id': task-1285135, 'name': PowerOnVM_Task, 'duration_secs': 0.700483} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.226718] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 902.226908] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-651032a3-dceb-41f0-af36-322221b8b9c5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Updating instance 'e6b0ce08-d67f-458f-92f5-1e904d03a5ef' progress to 100 {{(pid=63088) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 902.283707] env[63088]: DEBUG oslo_vmware.api [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52ce2e97-6cf5-51e3-7a17-5dc10a65a3d6, 'name': SearchDatastore_Task, 'duration_secs': 0.041182} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.283938] env[63088]: DEBUG oslo_concurrency.lockutils [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 902.284232] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] 71e37d8e-a454-46c4-a3cc-3d5671a32beb/71e37d8e-a454-46c4-a3cc-3d5671a32beb.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 902.284487] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3e8f8bc4-67d1-4240-a79c-b12f76ae2ca9 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.288667] env[63088]: DEBUG nova.network.neutron [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Port 1fb0e170-03ab-4ee4-9f1f-c5c6008c823d binding to destination host cpu-1 is already ACTIVE {{(pid=63088) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 902.291029] env[63088]: DEBUG oslo_vmware.api [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Waiting for the task: (returnval){ [ 902.291029] env[63088]: value = "task-1285139" [ 902.291029] env[63088]: _type = "Task" [ 902.291029] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.299159] env[63088]: DEBUG oslo_vmware.api [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285139, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.475639] env[63088]: DEBUG nova.scheduler.client.report [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 902.587713] env[63088]: DEBUG nova.compute.manager [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: bebc3318-24bb-4a37-8b23-66a12a7f7fd2] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 902.617352] env[63088]: DEBUG nova.virt.hardware [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 902.617783] env[63088]: DEBUG nova.virt.hardware [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 902.617988] env[63088]: DEBUG nova.virt.hardware [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 902.618230] env[63088]: DEBUG nova.virt.hardware [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 902.618475] env[63088]: DEBUG nova.virt.hardware [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 902.618769] env[63088]: DEBUG nova.virt.hardware [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 902.619048] env[63088]: DEBUG nova.virt.hardware [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 902.619275] env[63088]: DEBUG nova.virt.hardware [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 902.619498] env[63088]: DEBUG nova.virt.hardware [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 902.620053] env[63088]: DEBUG nova.virt.hardware [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 902.620272] env[63088]: DEBUG nova.virt.hardware [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 902.621500] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1d1f343-3a59-4b4b-b4c4-3443f9510bed {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.630875] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d916608-6c40-42c7-8aa5-94425e78170e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.647180] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-b81753c4-9726-4ca0-bfa8-881e9fef1e6f tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 254db932-35f1-42e3-9207-cd886efd65a3] Creating Snapshot of the VM instance {{(pid=63088) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 902.647750] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-03bb81dc-9727-4b90-8a21-a58c124afdf9 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.655615] env[63088]: DEBUG oslo_vmware.api [None req-b81753c4-9726-4ca0-bfa8-881e9fef1e6f tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Waiting for the task: (returnval){ [ 902.655615] env[63088]: value = "task-1285140" [ 902.655615] env[63088]: _type = "Task" [ 902.655615] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.666100] env[63088]: DEBUG oslo_vmware.api [None req-b81753c4-9726-4ca0-bfa8-881e9fef1e6f tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285140, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.806841] env[63088]: DEBUG oslo_vmware.api [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285139, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.980946] env[63088]: DEBUG oslo_concurrency.lockutils [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.408s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.981511] env[63088]: DEBUG nova.compute.manager [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 45513c6b-c000-4ee1-8893-4e084ffc22c5] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 902.984017] env[63088]: DEBUG oslo_concurrency.lockutils [None req-e757b386-7dab-43e3-b44f-3eab829d5a35 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.299s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.984251] env[63088]: DEBUG nova.objects.instance [None req-e757b386-7dab-43e3-b44f-3eab829d5a35 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Lazy-loading 'resources' on Instance uuid 58d561fe-26aa-4e94-8d55-cc70c361b479 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 903.100502] env[63088]: DEBUG nova.compute.manager [req-1021cefb-d14b-4b9e-b475-c3b08c4122fc req-b1158daf-ad9c-4fd2-827f-492db7c064ec service nova] [instance: bebc3318-24bb-4a37-8b23-66a12a7f7fd2] Received event network-vif-plugged-bd817fe9-5d27-40bf-9348-e29243b263f5 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 903.100720] env[63088]: DEBUG oslo_concurrency.lockutils [req-1021cefb-d14b-4b9e-b475-c3b08c4122fc req-b1158daf-ad9c-4fd2-827f-492db7c064ec service nova] Acquiring lock "bebc3318-24bb-4a37-8b23-66a12a7f7fd2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 903.100935] env[63088]: DEBUG oslo_concurrency.lockutils [req-1021cefb-d14b-4b9e-b475-c3b08c4122fc req-b1158daf-ad9c-4fd2-827f-492db7c064ec service nova] Lock "bebc3318-24bb-4a37-8b23-66a12a7f7fd2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 903.101126] env[63088]: DEBUG oslo_concurrency.lockutils [req-1021cefb-d14b-4b9e-b475-c3b08c4122fc req-b1158daf-ad9c-4fd2-827f-492db7c064ec service nova] Lock "bebc3318-24bb-4a37-8b23-66a12a7f7fd2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.101299] env[63088]: DEBUG nova.compute.manager [req-1021cefb-d14b-4b9e-b475-c3b08c4122fc req-b1158daf-ad9c-4fd2-827f-492db7c064ec service nova] [instance: bebc3318-24bb-4a37-8b23-66a12a7f7fd2] No waiting events found dispatching network-vif-plugged-bd817fe9-5d27-40bf-9348-e29243b263f5 {{(pid=63088) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 903.101466] env[63088]: WARNING nova.compute.manager [req-1021cefb-d14b-4b9e-b475-c3b08c4122fc req-b1158daf-ad9c-4fd2-827f-492db7c064ec service nova] [instance: bebc3318-24bb-4a37-8b23-66a12a7f7fd2] Received unexpected event network-vif-plugged-bd817fe9-5d27-40bf-9348-e29243b263f5 for instance with vm_state building and task_state spawning. [ 903.166639] env[63088]: DEBUG oslo_vmware.api [None req-b81753c4-9726-4ca0-bfa8-881e9fef1e6f tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285140, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.190839] env[63088]: DEBUG nova.network.neutron [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: bebc3318-24bb-4a37-8b23-66a12a7f7fd2] Successfully updated port: bd817fe9-5d27-40bf-9348-e29243b263f5 {{(pid=63088) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 903.313059] env[63088]: DEBUG oslo_vmware.api [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285139, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.529361} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.315014] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] 71e37d8e-a454-46c4-a3cc-3d5671a32beb/71e37d8e-a454-46c4-a3cc-3d5671a32beb.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 903.315248] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Extending root virtual disk to 1048576 {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 903.319163] env[63088]: DEBUG oslo_concurrency.lockutils [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquiring lock "2269e520-08b6-433f-8fe9-7b84d91e02d7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 903.319383] env[63088]: DEBUG oslo_concurrency.lockutils [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Lock "2269e520-08b6-433f-8fe9-7b84d91e02d7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 903.319557] env[63088]: DEBUG oslo_concurrency.lockutils [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Lock "2269e520-08b6-433f-8fe9-7b84d91e02d7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.320330] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-92533029-0a41-4030-8382-50ee853e0054 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.328146] env[63088]: DEBUG oslo_vmware.api [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Waiting for the task: (returnval){ [ 903.328146] env[63088]: value = "task-1285141" [ 903.328146] env[63088]: _type = "Task" [ 903.328146] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.337367] env[63088]: DEBUG oslo_vmware.api [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285141, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.487070] env[63088]: DEBUG nova.compute.utils [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 903.488999] env[63088]: DEBUG nova.compute.manager [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 45513c6b-c000-4ee1-8893-4e084ffc22c5] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 903.488999] env[63088]: DEBUG nova.network.neutron [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 45513c6b-c000-4ee1-8893-4e084ffc22c5] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 903.531799] env[63088]: DEBUG nova.policy [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '37d241cdf58e4aa7b05f1f24a3c78296', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e27be174c0b3400193f9ae64a242da1a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 903.669713] env[63088]: DEBUG oslo_vmware.api [None req-b81753c4-9726-4ca0-bfa8-881e9fef1e6f tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285140, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.693698] env[63088]: DEBUG oslo_concurrency.lockutils [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Acquiring lock "refresh_cache-bebc3318-24bb-4a37-8b23-66a12a7f7fd2" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 903.693835] env[63088]: DEBUG oslo_concurrency.lockutils [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Acquired lock "refresh_cache-bebc3318-24bb-4a37-8b23-66a12a7f7fd2" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 903.693988] env[63088]: DEBUG nova.network.neutron [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: bebc3318-24bb-4a37-8b23-66a12a7f7fd2] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 903.728627] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-703c14f5-0b58-461f-9f3e-89a8e81fb95e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.736442] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b193e3a2-5178-4221-993c-ae23e2a3f6c7 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.775055] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e875c7cf-c0b0-425d-92be-24f0e2e5fb88 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.785333] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d6d2435-ebf9-4605-a864-4b16d08f9843 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.801754] env[63088]: DEBUG nova.compute.provider_tree [None req-e757b386-7dab-43e3-b44f-3eab829d5a35 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 903.839949] env[63088]: DEBUG oslo_vmware.api [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285141, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.059179} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.840284] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Extended root virtual disk {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 903.841496] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29e89577-da8e-4612-a159-f3f3f6b2b0a6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.844802] env[63088]: DEBUG nova.network.neutron [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 45513c6b-c000-4ee1-8893-4e084ffc22c5] Successfully created port: 3d848d39-f98f-4614-918d-aba9f290becb {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 903.866770] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Reconfiguring VM instance instance-0000004d to attach disk [datastore1] 71e37d8e-a454-46c4-a3cc-3d5671a32beb/71e37d8e-a454-46c4-a3cc-3d5671a32beb.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 903.867106] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0867fa57-84f6-47cf-839e-cc52e55cc16b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.888638] env[63088]: DEBUG oslo_vmware.api [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Waiting for the task: (returnval){ [ 903.888638] env[63088]: value = "task-1285142" [ 903.888638] env[63088]: _type = "Task" [ 903.888638] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.902926] env[63088]: DEBUG oslo_vmware.api [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285142, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.993913] env[63088]: DEBUG nova.compute.manager [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 45513c6b-c000-4ee1-8893-4e084ffc22c5] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 904.170915] env[63088]: DEBUG oslo_vmware.api [None req-b81753c4-9726-4ca0-bfa8-881e9fef1e6f tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285140, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.236167] env[63088]: DEBUG nova.network.neutron [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: bebc3318-24bb-4a37-8b23-66a12a7f7fd2] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 904.306687] env[63088]: DEBUG nova.scheduler.client.report [None req-e757b386-7dab-43e3-b44f-3eab829d5a35 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 904.380683] env[63088]: DEBUG oslo_concurrency.lockutils [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquiring lock "refresh_cache-2269e520-08b6-433f-8fe9-7b84d91e02d7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 904.380953] env[63088]: DEBUG oslo_concurrency.lockutils [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquired lock "refresh_cache-2269e520-08b6-433f-8fe9-7b84d91e02d7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.381172] env[63088]: DEBUG nova.network.neutron [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 904.401123] env[63088]: DEBUG oslo_vmware.api [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285142, 'name': ReconfigVM_Task, 'duration_secs': 0.307296} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.403741] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Reconfigured VM instance instance-0000004d to attach disk [datastore1] 71e37d8e-a454-46c4-a3cc-3d5671a32beb/71e37d8e-a454-46c4-a3cc-3d5671a32beb.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 904.405225] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-da0a09ae-8625-47bd-8023-146cbe3a144c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.412828] env[63088]: DEBUG oslo_vmware.api [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Waiting for the task: (returnval){ [ 904.412828] env[63088]: value = "task-1285144" [ 904.412828] env[63088]: _type = "Task" [ 904.412828] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.420603] env[63088]: DEBUG oslo_vmware.api [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285144, 'name': Rename_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.430666] env[63088]: DEBUG nova.network.neutron [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: bebc3318-24bb-4a37-8b23-66a12a7f7fd2] Updating instance_info_cache with network_info: [{"id": "bd817fe9-5d27-40bf-9348-e29243b263f5", "address": "fa:16:3e:04:dd:3b", "network": {"id": "79155637-c844-4012-a77a-d04f9ec9ab18", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-2144942540-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e27be174c0b3400193f9ae64a242da1a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbd817fe9-5d", "ovs_interfaceid": "bd817fe9-5d27-40bf-9348-e29243b263f5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 904.491090] env[63088]: DEBUG nova.network.neutron [None req-1bc44084-4c8b-4bbf-b317-1818ae32cad5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Port 63d8dbbe-4550-4d8d-8e85-9a9eadc81d82 binding to destination host cpu-1 is already ACTIVE {{(pid=63088) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 904.491380] env[63088]: DEBUG oslo_concurrency.lockutils [None req-1bc44084-4c8b-4bbf-b317-1818ae32cad5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Acquiring lock "refresh_cache-e6b0ce08-d67f-458f-92f5-1e904d03a5ef" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 904.491537] env[63088]: DEBUG oslo_concurrency.lockutils [None req-1bc44084-4c8b-4bbf-b317-1818ae32cad5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Acquired lock "refresh_cache-e6b0ce08-d67f-458f-92f5-1e904d03a5ef" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.491705] env[63088]: DEBUG nova.network.neutron [None req-1bc44084-4c8b-4bbf-b317-1818ae32cad5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 904.670168] env[63088]: DEBUG oslo_vmware.api [None req-b81753c4-9726-4ca0-bfa8-881e9fef1e6f tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285140, 'name': CreateSnapshot_Task, 'duration_secs': 1.544382} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.670470] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-b81753c4-9726-4ca0-bfa8-881e9fef1e6f tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 254db932-35f1-42e3-9207-cd886efd65a3] Created Snapshot of the VM instance {{(pid=63088) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 904.671218] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2c2d44c-eb96-4dad-bf9b-f2baf5006f78 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.812335] env[63088]: DEBUG oslo_concurrency.lockutils [None req-e757b386-7dab-43e3-b44f-3eab829d5a35 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.828s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.814846] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f1666685-2914-415b-a4ae-382918967cb8 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.385s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.815154] env[63088]: DEBUG nova.objects.instance [None req-f1666685-2914-415b-a4ae-382918967cb8 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lazy-loading 'resources' on Instance uuid 0fdb21d1-4111-4ff3-bdc0-e2598298a9a4 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 904.836422] env[63088]: INFO nova.scheduler.client.report [None req-e757b386-7dab-43e3-b44f-3eab829d5a35 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Deleted allocations for instance 58d561fe-26aa-4e94-8d55-cc70c361b479 [ 904.921850] env[63088]: DEBUG oslo_vmware.api [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285144, 'name': Rename_Task, 'duration_secs': 0.132262} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.922079] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 904.922366] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ef8dcaa1-6187-4b71-a414-ecaac1df76bf {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.929970] env[63088]: DEBUG oslo_vmware.api [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Waiting for the task: (returnval){ [ 904.929970] env[63088]: value = "task-1285145" [ 904.929970] env[63088]: _type = "Task" [ 904.929970] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.935513] env[63088]: DEBUG oslo_concurrency.lockutils [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Releasing lock "refresh_cache-bebc3318-24bb-4a37-8b23-66a12a7f7fd2" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 904.935823] env[63088]: DEBUG nova.compute.manager [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: bebc3318-24bb-4a37-8b23-66a12a7f7fd2] Instance network_info: |[{"id": "bd817fe9-5d27-40bf-9348-e29243b263f5", "address": "fa:16:3e:04:dd:3b", "network": {"id": "79155637-c844-4012-a77a-d04f9ec9ab18", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-2144942540-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e27be174c0b3400193f9ae64a242da1a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbd817fe9-5d", "ovs_interfaceid": "bd817fe9-5d27-40bf-9348-e29243b263f5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 904.936224] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: bebc3318-24bb-4a37-8b23-66a12a7f7fd2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:04:dd:3b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '93c5b7ce-4c84-40bc-884c-b2453e0eee69', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bd817fe9-5d27-40bf-9348-e29243b263f5', 'vif_model': 'vmxnet3'}] {{(pid=63088) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 904.943700] env[63088]: DEBUG oslo.service.loopingcall [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 904.944377] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bebc3318-24bb-4a37-8b23-66a12a7f7fd2] Creating VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 904.944611] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f1ce8621-ee48-4d24-b946-4abc3a0135db {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.962902] env[63088]: DEBUG oslo_vmware.api [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285145, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.968491] env[63088]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 904.968491] env[63088]: value = "task-1285146" [ 904.968491] env[63088]: _type = "Task" [ 904.968491] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.976136] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285146, 'name': CreateVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.004298] env[63088]: DEBUG nova.compute.manager [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 45513c6b-c000-4ee1-8893-4e084ffc22c5] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 905.035410] env[63088]: DEBUG nova.virt.hardware [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 905.035754] env[63088]: DEBUG nova.virt.hardware [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 905.035967] env[63088]: DEBUG nova.virt.hardware [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 905.036245] env[63088]: DEBUG nova.virt.hardware [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 905.036453] env[63088]: DEBUG nova.virt.hardware [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 905.036660] env[63088]: DEBUG nova.virt.hardware [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 905.037082] env[63088]: DEBUG nova.virt.hardware [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 905.037191] env[63088]: DEBUG nova.virt.hardware [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 905.037398] env[63088]: DEBUG nova.virt.hardware [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 905.037621] env[63088]: DEBUG nova.virt.hardware [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 905.037858] env[63088]: DEBUG nova.virt.hardware [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 905.039088] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f2ffe76-421d-4729-bfde-1849645ed054 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.059082] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-006c1543-e1cb-4f21-b6fa-f38fbba0bef6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.174403] env[63088]: DEBUG nova.compute.manager [req-17646c1a-b067-4b98-ad97-2be36642651a req-6b08ffa1-5823-4316-aa7a-a3e712ac990c service nova] [instance: bebc3318-24bb-4a37-8b23-66a12a7f7fd2] Received event network-changed-bd817fe9-5d27-40bf-9348-e29243b263f5 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 905.176449] env[63088]: DEBUG nova.compute.manager [req-17646c1a-b067-4b98-ad97-2be36642651a req-6b08ffa1-5823-4316-aa7a-a3e712ac990c service nova] [instance: bebc3318-24bb-4a37-8b23-66a12a7f7fd2] Refreshing instance network info cache due to event network-changed-bd817fe9-5d27-40bf-9348-e29243b263f5. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 905.176449] env[63088]: DEBUG oslo_concurrency.lockutils [req-17646c1a-b067-4b98-ad97-2be36642651a req-6b08ffa1-5823-4316-aa7a-a3e712ac990c service nova] Acquiring lock "refresh_cache-bebc3318-24bb-4a37-8b23-66a12a7f7fd2" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 905.176449] env[63088]: DEBUG oslo_concurrency.lockutils [req-17646c1a-b067-4b98-ad97-2be36642651a req-6b08ffa1-5823-4316-aa7a-a3e712ac990c service nova] Acquired lock "refresh_cache-bebc3318-24bb-4a37-8b23-66a12a7f7fd2" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 905.176449] env[63088]: DEBUG nova.network.neutron [req-17646c1a-b067-4b98-ad97-2be36642651a req-6b08ffa1-5823-4316-aa7a-a3e712ac990c service nova] [instance: bebc3318-24bb-4a37-8b23-66a12a7f7fd2] Refreshing network info cache for port bd817fe9-5d27-40bf-9348-e29243b263f5 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 905.179476] env[63088]: DEBUG nova.network.neutron [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Updating instance_info_cache with network_info: [{"id": "1fb0e170-03ab-4ee4-9f1f-c5c6008c823d", "address": "fa:16:3e:a8:02:36", "network": {"id": "dff14a7f-0af1-4e4e-a498-86d7c9816e6b", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-762073671-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f33f2701fad94864a8c406a404bc0a42", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "21310d90-efbc-45a8-a97f-c4358606530f", "external-id": "nsx-vlan-transportzone-672", "segmentation_id": 672, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1fb0e170-03", "ovs_interfaceid": "1fb0e170-03ab-4ee4-9f1f-c5c6008c823d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 905.191620] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-b81753c4-9726-4ca0-bfa8-881e9fef1e6f tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 254db932-35f1-42e3-9207-cd886efd65a3] Creating linked-clone VM from snapshot {{(pid=63088) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 905.193583] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-bd6afdfc-df66-466a-80a4-36e436f3ab96 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.204194] env[63088]: DEBUG oslo_vmware.api [None req-b81753c4-9726-4ca0-bfa8-881e9fef1e6f tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Waiting for the task: (returnval){ [ 905.204194] env[63088]: value = "task-1285147" [ 905.204194] env[63088]: _type = "Task" [ 905.204194] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.214208] env[63088]: DEBUG oslo_vmware.api [None req-b81753c4-9726-4ca0-bfa8-881e9fef1e6f tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285147, 'name': CloneVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.349074] env[63088]: DEBUG oslo_concurrency.lockutils [None req-e757b386-7dab-43e3-b44f-3eab829d5a35 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Lock "58d561fe-26aa-4e94-8d55-cc70c361b479" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.527s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 905.408926] env[63088]: DEBUG nova.network.neutron [None req-1bc44084-4c8b-4bbf-b317-1818ae32cad5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Updating instance_info_cache with network_info: [{"id": "63d8dbbe-4550-4d8d-8e85-9a9eadc81d82", "address": "fa:16:3e:d1:72:56", "network": {"id": "ebeefbc2-043c-4e76-aeaf-4f9113869855", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.60", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "bfb6bc5062fc4d51af1d3c577659be9a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271fe7a0-dfd7-409b-920a-cf83ef1a86a3", "external-id": "nsx-vlan-transportzone-728", "segmentation_id": 728, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap63d8dbbe-45", "ovs_interfaceid": "63d8dbbe-4550-4d8d-8e85-9a9eadc81d82", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 905.440065] env[63088]: DEBUG oslo_vmware.api [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285145, 'name': PowerOnVM_Task, 'duration_secs': 0.470208} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.442709] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 905.442923] env[63088]: INFO nova.compute.manager [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Took 7.76 seconds to spawn the instance on the hypervisor. [ 905.443120] env[63088]: DEBUG nova.compute.manager [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 905.444479] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04a1c78a-5295-4725-a4c9-a27f09d2ee09 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.481082] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285146, 'name': CreateVM_Task, 'duration_secs': 0.393343} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.481082] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bebc3318-24bb-4a37-8b23-66a12a7f7fd2] Created VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 905.481222] env[63088]: DEBUG oslo_concurrency.lockutils [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 905.482472] env[63088]: DEBUG oslo_concurrency.lockutils [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 905.482472] env[63088]: DEBUG oslo_concurrency.lockutils [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 905.482472] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c50493eb-4850-410f-9729-19593ca2d753 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.489414] env[63088]: DEBUG oslo_vmware.api [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Waiting for the task: (returnval){ [ 905.489414] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]5275e9f3-ff72-3da9-28ad-5c5794dae167" [ 905.489414] env[63088]: _type = "Task" [ 905.489414] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.499424] env[63088]: DEBUG oslo_vmware.api [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]5275e9f3-ff72-3da9-28ad-5c5794dae167, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.601348] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86601d51-a517-4cd1-9e5d-79f547faa84e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.609389] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e974f976-5446-4b63-ba3e-2147221cb35b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.642184] env[63088]: DEBUG nova.network.neutron [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 45513c6b-c000-4ee1-8893-4e084ffc22c5] Successfully updated port: 3d848d39-f98f-4614-918d-aba9f290becb {{(pid=63088) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 905.643778] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ced530c-fcda-4077-9cdc-5ff067ebbc42 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.646537] env[63088]: DEBUG oslo_concurrency.lockutils [None req-2bccf470-f018-4694-b367-d6d0046e5e91 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Acquiring lock "db032da3-d334-486f-a428-af1c8a3c360a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 905.646766] env[63088]: DEBUG oslo_concurrency.lockutils [None req-2bccf470-f018-4694-b367-d6d0046e5e91 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Lock "db032da3-d334-486f-a428-af1c8a3c360a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 905.646968] env[63088]: DEBUG oslo_concurrency.lockutils [None req-2bccf470-f018-4694-b367-d6d0046e5e91 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Acquiring lock "db032da3-d334-486f-a428-af1c8a3c360a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 905.647182] env[63088]: DEBUG oslo_concurrency.lockutils [None req-2bccf470-f018-4694-b367-d6d0046e5e91 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Lock "db032da3-d334-486f-a428-af1c8a3c360a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 905.647352] env[63088]: DEBUG oslo_concurrency.lockutils [None req-2bccf470-f018-4694-b367-d6d0046e5e91 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Lock "db032da3-d334-486f-a428-af1c8a3c360a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 905.650166] env[63088]: INFO nova.compute.manager [None req-2bccf470-f018-4694-b367-d6d0046e5e91 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: db032da3-d334-486f-a428-af1c8a3c360a] Terminating instance [ 905.654160] env[63088]: DEBUG nova.compute.manager [None req-2bccf470-f018-4694-b367-d6d0046e5e91 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: db032da3-d334-486f-a428-af1c8a3c360a] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 905.654365] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-2bccf470-f018-4694-b367-d6d0046e5e91 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: db032da3-d334-486f-a428-af1c8a3c360a] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 905.655557] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8c30cb3-fb30-4fb2-90e9-c28e61f7c5c4 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.659909] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ac62b26-94f4-4041-9814-4d3285c5d3bf {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.673527] env[63088]: DEBUG nova.compute.provider_tree [None req-f1666685-2914-415b-a4ae-382918967cb8 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 905.676335] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-2bccf470-f018-4694-b367-d6d0046e5e91 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: db032da3-d334-486f-a428-af1c8a3c360a] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 905.676676] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-89c7ee6a-5229-4403-811d-4bba17cdddcf {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.681112] env[63088]: DEBUG oslo_concurrency.lockutils [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Releasing lock "refresh_cache-2269e520-08b6-433f-8fe9-7b84d91e02d7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 905.685550] env[63088]: DEBUG oslo_vmware.api [None req-2bccf470-f018-4694-b367-d6d0046e5e91 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Waiting for the task: (returnval){ [ 905.685550] env[63088]: value = "task-1285148" [ 905.685550] env[63088]: _type = "Task" [ 905.685550] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.693980] env[63088]: DEBUG oslo_vmware.api [None req-2bccf470-f018-4694-b367-d6d0046e5e91 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': task-1285148, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.715998] env[63088]: DEBUG oslo_vmware.api [None req-b81753c4-9726-4ca0-bfa8-881e9fef1e6f tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285147, 'name': CloneVM_Task} progress is 94%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.885798] env[63088]: DEBUG nova.network.neutron [req-17646c1a-b067-4b98-ad97-2be36642651a req-6b08ffa1-5823-4316-aa7a-a3e712ac990c service nova] [instance: bebc3318-24bb-4a37-8b23-66a12a7f7fd2] Updated VIF entry in instance network info cache for port bd817fe9-5d27-40bf-9348-e29243b263f5. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 905.886221] env[63088]: DEBUG nova.network.neutron [req-17646c1a-b067-4b98-ad97-2be36642651a req-6b08ffa1-5823-4316-aa7a-a3e712ac990c service nova] [instance: bebc3318-24bb-4a37-8b23-66a12a7f7fd2] Updating instance_info_cache with network_info: [{"id": "bd817fe9-5d27-40bf-9348-e29243b263f5", "address": "fa:16:3e:04:dd:3b", "network": {"id": "79155637-c844-4012-a77a-d04f9ec9ab18", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-2144942540-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e27be174c0b3400193f9ae64a242da1a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbd817fe9-5d", "ovs_interfaceid": "bd817fe9-5d27-40bf-9348-e29243b263f5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 905.915057] env[63088]: DEBUG oslo_concurrency.lockutils [None req-1bc44084-4c8b-4bbf-b317-1818ae32cad5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Releasing lock "refresh_cache-e6b0ce08-d67f-458f-92f5-1e904d03a5ef" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 905.961603] env[63088]: INFO nova.compute.manager [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Took 27.78 seconds to build instance. [ 906.000523] env[63088]: DEBUG oslo_vmware.api [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]5275e9f3-ff72-3da9-28ad-5c5794dae167, 'name': SearchDatastore_Task, 'duration_secs': 0.011573} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.000839] env[63088]: DEBUG oslo_concurrency.lockutils [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 906.001130] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: bebc3318-24bb-4a37-8b23-66a12a7f7fd2] Processing image 1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 906.001405] env[63088]: DEBUG oslo_concurrency.lockutils [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 906.001569] env[63088]: DEBUG oslo_concurrency.lockutils [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 906.001758] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 906.002026] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-24ea8604-749e-4c84-8c92-aacd1f27bfbf {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.010664] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 906.010859] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 906.011617] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b2b24017-9125-434c-a281-da9144177ab3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.016778] env[63088]: DEBUG oslo_vmware.api [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Waiting for the task: (returnval){ [ 906.016778] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]5264663e-0742-caae-99b1-4c1df9a127d7" [ 906.016778] env[63088]: _type = "Task" [ 906.016778] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.024306] env[63088]: DEBUG oslo_vmware.api [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]5264663e-0742-caae-99b1-4c1df9a127d7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.147706] env[63088]: DEBUG oslo_concurrency.lockutils [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Acquiring lock "refresh_cache-45513c6b-c000-4ee1-8893-4e084ffc22c5" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 906.147874] env[63088]: DEBUG oslo_concurrency.lockutils [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Acquired lock "refresh_cache-45513c6b-c000-4ee1-8893-4e084ffc22c5" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 906.148065] env[63088]: DEBUG nova.network.neutron [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 45513c6b-c000-4ee1-8893-4e084ffc22c5] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 906.178272] env[63088]: DEBUG nova.scheduler.client.report [None req-f1666685-2914-415b-a4ae-382918967cb8 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 906.202395] env[63088]: DEBUG oslo_vmware.api [None req-2bccf470-f018-4694-b367-d6d0046e5e91 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': task-1285148, 'name': PowerOffVM_Task, 'duration_secs': 0.216328} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.202691] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-2bccf470-f018-4694-b367-d6d0046e5e91 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: db032da3-d334-486f-a428-af1c8a3c360a] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 906.202863] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-2bccf470-f018-4694-b367-d6d0046e5e91 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: db032da3-d334-486f-a428-af1c8a3c360a] Unregistering the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 906.203198] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-22d0efe6-55e9-4cbe-b19a-e5a8e8e8c243 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.209687] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9e71018-330f-4c91-85cf-baf81d690633 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.217855] env[63088]: DEBUG oslo_vmware.api [None req-b81753c4-9726-4ca0-bfa8-881e9fef1e6f tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285147, 'name': CloneVM_Task} progress is 94%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.232770] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd6c6a4a-ee10-4404-a365-feb64d291ea1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.240480] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Updating instance '2269e520-08b6-433f-8fe9-7b84d91e02d7' progress to 83 {{(pid=63088) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 906.291851] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-2bccf470-f018-4694-b367-d6d0046e5e91 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: db032da3-d334-486f-a428-af1c8a3c360a] Unregistered the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 906.292175] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-2bccf470-f018-4694-b367-d6d0046e5e91 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: db032da3-d334-486f-a428-af1c8a3c360a] Deleting contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 906.292384] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-2bccf470-f018-4694-b367-d6d0046e5e91 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Deleting the datastore file [datastore1] db032da3-d334-486f-a428-af1c8a3c360a {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 906.292697] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6a69e6bc-4dbe-46f9-8195-aa684017d7f6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.299310] env[63088]: DEBUG oslo_vmware.api [None req-2bccf470-f018-4694-b367-d6d0046e5e91 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Waiting for the task: (returnval){ [ 906.299310] env[63088]: value = "task-1285150" [ 906.299310] env[63088]: _type = "Task" [ 906.299310] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.308853] env[63088]: DEBUG oslo_vmware.api [None req-2bccf470-f018-4694-b367-d6d0046e5e91 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': task-1285150, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.388680] env[63088]: DEBUG oslo_concurrency.lockutils [req-17646c1a-b067-4b98-ad97-2be36642651a req-6b08ffa1-5823-4316-aa7a-a3e712ac990c service nova] Releasing lock "refresh_cache-bebc3318-24bb-4a37-8b23-66a12a7f7fd2" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 906.418203] env[63088]: DEBUG nova.compute.manager [None req-1bc44084-4c8b-4bbf-b317-1818ae32cad5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=63088) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:897}} [ 906.418453] env[63088]: DEBUG oslo_concurrency.lockutils [None req-1bc44084-4c8b-4bbf-b317-1818ae32cad5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.463920] env[63088]: DEBUG oslo_concurrency.lockutils [None req-40f1454b-a90e-4337-a097-ab0f98a13a2d tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Lock "71e37d8e-a454-46c4-a3cc-3d5671a32beb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.289s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.527375] env[63088]: DEBUG oslo_vmware.api [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]5264663e-0742-caae-99b1-4c1df9a127d7, 'name': SearchDatastore_Task, 'duration_secs': 0.01035} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.528309] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-82cb0f2c-ad6d-43f4-852c-db2651cde07e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.533694] env[63088]: DEBUG oslo_vmware.api [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Waiting for the task: (returnval){ [ 906.533694] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52468afc-859a-9a96-e123-62a3ab8f53ac" [ 906.533694] env[63088]: _type = "Task" [ 906.533694] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.541694] env[63088]: DEBUG oslo_vmware.api [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52468afc-859a-9a96-e123-62a3ab8f53ac, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.555568] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-e2dee0e1-16ef-4ce0-9a3b-bc5d82868451 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Volume attach. Driver type: vmdk {{(pid=63088) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 906.555851] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-e2dee0e1-16ef-4ce0-9a3b-bc5d82868451 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-275902', 'volume_id': 'deedf893-eb6c-4ff0-ab0b-3e34abe88c8f', 'name': 'volume-deedf893-eb6c-4ff0-ab0b-3e34abe88c8f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69', 'attached_at': '', 'detached_at': '', 'volume_id': 'deedf893-eb6c-4ff0-ab0b-3e34abe88c8f', 'serial': 'deedf893-eb6c-4ff0-ab0b-3e34abe88c8f'} {{(pid=63088) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 906.556825] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d9abf2d-1b57-4412-a349-ff96c7b51a1b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.573991] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-303053d0-8304-41b4-b3fd-4ae35262ad5f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.602906] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-e2dee0e1-16ef-4ce0-9a3b-bc5d82868451 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Reconfiguring VM instance instance-00000043 to attach disk [datastore1] volume-deedf893-eb6c-4ff0-ab0b-3e34abe88c8f/volume-deedf893-eb6c-4ff0-ab0b-3e34abe88c8f.vmdk or device None with type thin {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 906.602906] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-eef0989e-39f0-489b-aa96-2bf1814c383a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.622228] env[63088]: DEBUG oslo_vmware.api [None req-e2dee0e1-16ef-4ce0-9a3b-bc5d82868451 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Waiting for the task: (returnval){ [ 906.622228] env[63088]: value = "task-1285151" [ 906.622228] env[63088]: _type = "Task" [ 906.622228] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.630088] env[63088]: DEBUG oslo_vmware.api [None req-e2dee0e1-16ef-4ce0-9a3b-bc5d82868451 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Task: {'id': task-1285151, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.683949] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f1666685-2914-415b-a4ae-382918967cb8 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.869s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.687091] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 19.923s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.700437] env[63088]: DEBUG nova.network.neutron [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 45513c6b-c000-4ee1-8893-4e084ffc22c5] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 906.706525] env[63088]: INFO nova.scheduler.client.report [None req-f1666685-2914-415b-a4ae-382918967cb8 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Deleted allocations for instance 0fdb21d1-4111-4ff3-bdc0-e2598298a9a4 [ 906.723023] env[63088]: DEBUG oslo_vmware.api [None req-b81753c4-9726-4ca0-bfa8-881e9fef1e6f tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285147, 'name': CloneVM_Task} progress is 94%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.749095] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 906.749095] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7886925a-0b73-4ae5-8628-5d4a406fa0c6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.757340] env[63088]: DEBUG oslo_vmware.api [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Waiting for the task: (returnval){ [ 906.757340] env[63088]: value = "task-1285152" [ 906.757340] env[63088]: _type = "Task" [ 906.757340] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.768916] env[63088]: DEBUG oslo_vmware.api [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285152, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.811382] env[63088]: DEBUG oslo_vmware.api [None req-2bccf470-f018-4694-b367-d6d0046e5e91 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Task: {'id': task-1285150, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.327923} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.811839] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-2bccf470-f018-4694-b367-d6d0046e5e91 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 906.812261] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-2bccf470-f018-4694-b367-d6d0046e5e91 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: db032da3-d334-486f-a428-af1c8a3c360a] Deleted contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 906.812596] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-2bccf470-f018-4694-b367-d6d0046e5e91 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: db032da3-d334-486f-a428-af1c8a3c360a] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 906.812934] env[63088]: INFO nova.compute.manager [None req-2bccf470-f018-4694-b367-d6d0046e5e91 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] [instance: db032da3-d334-486f-a428-af1c8a3c360a] Took 1.16 seconds to destroy the instance on the hypervisor. [ 906.813380] env[63088]: DEBUG oslo.service.loopingcall [None req-2bccf470-f018-4694-b367-d6d0046e5e91 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 906.813740] env[63088]: DEBUG nova.compute.manager [-] [instance: db032da3-d334-486f-a428-af1c8a3c360a] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 906.813965] env[63088]: DEBUG nova.network.neutron [-] [instance: db032da3-d334-486f-a428-af1c8a3c360a] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 907.047271] env[63088]: DEBUG oslo_vmware.api [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52468afc-859a-9a96-e123-62a3ab8f53ac, 'name': SearchDatastore_Task, 'duration_secs': 0.009133} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.048232] env[63088]: DEBUG nova.network.neutron [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 45513c6b-c000-4ee1-8893-4e084ffc22c5] Updating instance_info_cache with network_info: [{"id": "3d848d39-f98f-4614-918d-aba9f290becb", "address": "fa:16:3e:08:54:08", "network": {"id": "79155637-c844-4012-a77a-d04f9ec9ab18", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-2144942540-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e27be174c0b3400193f9ae64a242da1a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d848d39-f9", "ovs_interfaceid": "3d848d39-f98f-4614-918d-aba9f290becb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 907.049506] env[63088]: DEBUG oslo_concurrency.lockutils [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 907.049778] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] bebc3318-24bb-4a37-8b23-66a12a7f7fd2/bebc3318-24bb-4a37-8b23-66a12a7f7fd2.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 907.050096] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c81f0f86-6b40-4445-861c-dc1cceb4cd20 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.057266] env[63088]: DEBUG oslo_vmware.api [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Waiting for the task: (returnval){ [ 907.057266] env[63088]: value = "task-1285153" [ 907.057266] env[63088]: _type = "Task" [ 907.057266] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.065843] env[63088]: DEBUG oslo_vmware.api [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': task-1285153, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.132659] env[63088]: DEBUG oslo_vmware.api [None req-e2dee0e1-16ef-4ce0-9a3b-bc5d82868451 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Task: {'id': task-1285151, 'name': ReconfigVM_Task, 'duration_secs': 0.466153} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.132946] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-e2dee0e1-16ef-4ce0-9a3b-bc5d82868451 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Reconfigured VM instance instance-00000043 to attach disk [datastore1] volume-deedf893-eb6c-4ff0-ab0b-3e34abe88c8f/volume-deedf893-eb6c-4ff0-ab0b-3e34abe88c8f.vmdk or device None with type thin {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 907.137974] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ed291523-3ce2-4477-9c8f-6540abcd7ad5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.155235] env[63088]: DEBUG oslo_vmware.api [None req-e2dee0e1-16ef-4ce0-9a3b-bc5d82868451 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Waiting for the task: (returnval){ [ 907.155235] env[63088]: value = "task-1285154" [ 907.155235] env[63088]: _type = "Task" [ 907.155235] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.167586] env[63088]: DEBUG oslo_vmware.api [None req-e2dee0e1-16ef-4ce0-9a3b-bc5d82868451 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Task: {'id': task-1285154, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.220102] env[63088]: DEBUG oslo_vmware.api [None req-b81753c4-9726-4ca0-bfa8-881e9fef1e6f tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285147, 'name': CloneVM_Task, 'duration_secs': 1.932602} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.221522] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f1666685-2914-415b-a4ae-382918967cb8 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lock "0fdb21d1-4111-4ff3-bdc0-e2598298a9a4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.013s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 907.222388] env[63088]: INFO nova.virt.vmwareapi.vmops [None req-b81753c4-9726-4ca0-bfa8-881e9fef1e6f tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 254db932-35f1-42e3-9207-cd886efd65a3] Created linked-clone VM from snapshot [ 907.223515] env[63088]: DEBUG nova.compute.manager [req-eef8ce2c-114a-401e-a4a3-a4405a370a29 req-45026a6b-b78b-4b4d-abe7-d52eb5454b3f service nova] [instance: 45513c6b-c000-4ee1-8893-4e084ffc22c5] Received event network-vif-plugged-3d848d39-f98f-4614-918d-aba9f290becb {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 907.223725] env[63088]: DEBUG oslo_concurrency.lockutils [req-eef8ce2c-114a-401e-a4a3-a4405a370a29 req-45026a6b-b78b-4b4d-abe7-d52eb5454b3f service nova] Acquiring lock "45513c6b-c000-4ee1-8893-4e084ffc22c5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 907.223928] env[63088]: DEBUG oslo_concurrency.lockutils [req-eef8ce2c-114a-401e-a4a3-a4405a370a29 req-45026a6b-b78b-4b4d-abe7-d52eb5454b3f service nova] Lock "45513c6b-c000-4ee1-8893-4e084ffc22c5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 907.224108] env[63088]: DEBUG oslo_concurrency.lockutils [req-eef8ce2c-114a-401e-a4a3-a4405a370a29 req-45026a6b-b78b-4b4d-abe7-d52eb5454b3f service nova] Lock "45513c6b-c000-4ee1-8893-4e084ffc22c5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 907.224285] env[63088]: DEBUG nova.compute.manager [req-eef8ce2c-114a-401e-a4a3-a4405a370a29 req-45026a6b-b78b-4b4d-abe7-d52eb5454b3f service nova] [instance: 45513c6b-c000-4ee1-8893-4e084ffc22c5] No waiting events found dispatching network-vif-plugged-3d848d39-f98f-4614-918d-aba9f290becb {{(pid=63088) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 907.224449] env[63088]: WARNING nova.compute.manager [req-eef8ce2c-114a-401e-a4a3-a4405a370a29 req-45026a6b-b78b-4b4d-abe7-d52eb5454b3f service nova] [instance: 45513c6b-c000-4ee1-8893-4e084ffc22c5] Received unexpected event network-vif-plugged-3d848d39-f98f-4614-918d-aba9f290becb for instance with vm_state building and task_state spawning. [ 907.224611] env[63088]: DEBUG nova.compute.manager [req-eef8ce2c-114a-401e-a4a3-a4405a370a29 req-45026a6b-b78b-4b4d-abe7-d52eb5454b3f service nova] [instance: 45513c6b-c000-4ee1-8893-4e084ffc22c5] Received event network-changed-3d848d39-f98f-4614-918d-aba9f290becb {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 907.224766] env[63088]: DEBUG nova.compute.manager [req-eef8ce2c-114a-401e-a4a3-a4405a370a29 req-45026a6b-b78b-4b4d-abe7-d52eb5454b3f service nova] [instance: 45513c6b-c000-4ee1-8893-4e084ffc22c5] Refreshing instance network info cache due to event network-changed-3d848d39-f98f-4614-918d-aba9f290becb. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 907.224935] env[63088]: DEBUG oslo_concurrency.lockutils [req-eef8ce2c-114a-401e-a4a3-a4405a370a29 req-45026a6b-b78b-4b4d-abe7-d52eb5454b3f service nova] Acquiring lock "refresh_cache-45513c6b-c000-4ee1-8893-4e084ffc22c5" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 907.226533] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc60b9c9-7a4b-4c49-86a8-942a0621dca3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.235724] env[63088]: DEBUG nova.virt.vmwareapi.images [None req-b81753c4-9726-4ca0-bfa8-881e9fef1e6f tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 254db932-35f1-42e3-9207-cd886efd65a3] Uploading image c89225f6-2d34-4988-ae47-d3d6903890da {{(pid=63088) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 907.272520] env[63088]: DEBUG oslo_vmware.api [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285152, 'name': PowerOnVM_Task, 'duration_secs': 0.425137} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.275135] env[63088]: DEBUG oslo_vmware.rw_handles [None req-b81753c4-9726-4ca0-bfa8-881e9fef1e6f tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 907.275135] env[63088]: value = "vm-275905" [ 907.275135] env[63088]: _type = "VirtualMachine" [ 907.275135] env[63088]: }. {{(pid=63088) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 907.275498] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 907.275683] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-df0af2b7-17a1-4220-b4da-b128f71800f0 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Updating instance '2269e520-08b6-433f-8fe9-7b84d91e02d7' progress to 100 {{(pid=63088) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 907.279384] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-a885dde6-4fdd-4edd-9d1d-512d96f18d50 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.287073] env[63088]: DEBUG oslo_vmware.rw_handles [None req-b81753c4-9726-4ca0-bfa8-881e9fef1e6f tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Lease: (returnval){ [ 907.287073] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52301661-3b30-c310-240c-271a50b2dabe" [ 907.287073] env[63088]: _type = "HttpNfcLease" [ 907.287073] env[63088]: } obtained for exporting VM: (result){ [ 907.287073] env[63088]: value = "vm-275905" [ 907.287073] env[63088]: _type = "VirtualMachine" [ 907.287073] env[63088]: }. {{(pid=63088) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 907.287319] env[63088]: DEBUG oslo_vmware.api [None req-b81753c4-9726-4ca0-bfa8-881e9fef1e6f tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Waiting for the lease: (returnval){ [ 907.287319] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52301661-3b30-c310-240c-271a50b2dabe" [ 907.287319] env[63088]: _type = "HttpNfcLease" [ 907.287319] env[63088]: } to be ready. {{(pid=63088) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 907.295542] env[63088]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 907.295542] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52301661-3b30-c310-240c-271a50b2dabe" [ 907.295542] env[63088]: _type = "HttpNfcLease" [ 907.295542] env[63088]: } is initializing. {{(pid=63088) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 907.554194] env[63088]: DEBUG oslo_concurrency.lockutils [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Releasing lock "refresh_cache-45513c6b-c000-4ee1-8893-4e084ffc22c5" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 907.554709] env[63088]: DEBUG nova.compute.manager [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 45513c6b-c000-4ee1-8893-4e084ffc22c5] Instance network_info: |[{"id": "3d848d39-f98f-4614-918d-aba9f290becb", "address": "fa:16:3e:08:54:08", "network": {"id": "79155637-c844-4012-a77a-d04f9ec9ab18", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-2144942540-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e27be174c0b3400193f9ae64a242da1a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d848d39-f9", "ovs_interfaceid": "3d848d39-f98f-4614-918d-aba9f290becb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 907.555244] env[63088]: DEBUG oslo_concurrency.lockutils [req-eef8ce2c-114a-401e-a4a3-a4405a370a29 req-45026a6b-b78b-4b4d-abe7-d52eb5454b3f service nova] Acquired lock "refresh_cache-45513c6b-c000-4ee1-8893-4e084ffc22c5" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 907.555568] env[63088]: DEBUG nova.network.neutron [req-eef8ce2c-114a-401e-a4a3-a4405a370a29 req-45026a6b-b78b-4b4d-abe7-d52eb5454b3f service nova] [instance: 45513c6b-c000-4ee1-8893-4e084ffc22c5] Refreshing network info cache for port 3d848d39-f98f-4614-918d-aba9f290becb {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 907.561772] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 45513c6b-c000-4ee1-8893-4e084ffc22c5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:08:54:08', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '93c5b7ce-4c84-40bc-884c-b2453e0eee69', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3d848d39-f98f-4614-918d-aba9f290becb', 'vif_model': 'vmxnet3'}] {{(pid=63088) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 907.579149] env[63088]: DEBUG oslo.service.loopingcall [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 907.582020] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 45513c6b-c000-4ee1-8893-4e084ffc22c5] Creating VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 907.586534] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-98536510-ef93-4223-bc71-a7f96b845aa3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.606122] env[63088]: DEBUG oslo_vmware.api [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': task-1285153, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.608082] env[63088]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 907.608082] env[63088]: value = "task-1285156" [ 907.608082] env[63088]: _type = "Task" [ 907.608082] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.615718] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285156, 'name': CreateVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.668169] env[63088]: DEBUG oslo_vmware.api [None req-e2dee0e1-16ef-4ce0-9a3b-bc5d82868451 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Task: {'id': task-1285154, 'name': ReconfigVM_Task, 'duration_secs': 0.158857} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.668640] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-e2dee0e1-16ef-4ce0-9a3b-bc5d82868451 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-275902', 'volume_id': 'deedf893-eb6c-4ff0-ab0b-3e34abe88c8f', 'name': 'volume-deedf893-eb6c-4ff0-ab0b-3e34abe88c8f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69', 'attached_at': '', 'detached_at': '', 'volume_id': 'deedf893-eb6c-4ff0-ab0b-3e34abe88c8f', 'serial': 'deedf893-eb6c-4ff0-ab0b-3e34abe88c8f'} {{(pid=63088) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 907.704558] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Applying migration context for instance 2269e520-08b6-433f-8fe9-7b84d91e02d7 as it has an incoming, in-progress migration 5602a4d9-9018-483a-98f7-e1e412ff902a. Migration status is post-migrating {{(pid=63088) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 907.704558] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Applying migration context for instance e6b0ce08-d67f-458f-92f5-1e904d03a5ef as it has an incoming, in-progress migration 02e19d97-e38c-43e6-b757-056d827d87dc. Migration status is reverting {{(pid=63088) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 907.706818] env[63088]: INFO nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Updating resource usage from migration 5602a4d9-9018-483a-98f7-e1e412ff902a [ 907.707089] env[63088]: INFO nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Updating resource usage from migration 02e19d97-e38c-43e6-b757-056d827d87dc [ 907.730309] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance db032da3-d334-486f-a428-af1c8a3c360a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 907.730522] env[63088]: WARNING nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance 4971b24c-6710-4f50-9846-727dad264b1f is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 907.730653] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance 7b6aadb7-e34b-42b7-b69f-370434f5b665 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 907.730773] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 907.730903] env[63088]: WARNING nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance 9d5f1cde-e787-4a2c-ac49-83ec135ff51c is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 907.731127] env[63088]: WARNING nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance 667fd5e9-5fe4-41e1-9d8a-896c6e5c6286 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 907.731251] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance b50db882-598c-488f-b935-34f55f655642 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 907.731364] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance 2964db35-7357-40a7-b4e6-7e2595549f5b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 907.731475] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Migration 02e19d97-e38c-43e6-b757-056d827d87dc is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 907.731585] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance e6b0ce08-d67f-458f-92f5-1e904d03a5ef actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 907.731692] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance 254db932-35f1-42e3-9207-cd886efd65a3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 907.731800] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Migration 5602a4d9-9018-483a-98f7-e1e412ff902a is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 907.731908] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance 2269e520-08b6-433f-8fe9-7b84d91e02d7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 907.732033] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance 71e37d8e-a454-46c4-a3cc-3d5671a32beb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 907.732150] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance bebc3318-24bb-4a37-8b23-66a12a7f7fd2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 907.732260] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance 45513c6b-c000-4ee1-8893-4e084ffc22c5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 907.732485] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Total usable vcpus: 48, total allocated vcpus: 13 {{(pid=63088) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 907.732653] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3136MB phys_disk=200GB used_disk=13GB total_vcpus=48 used_vcpus=13 pci_stats=[] {{(pid=63088) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 907.795335] env[63088]: DEBUG nova.network.neutron [-] [instance: db032da3-d334-486f-a428-af1c8a3c360a] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 907.799735] env[63088]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 907.799735] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52301661-3b30-c310-240c-271a50b2dabe" [ 907.799735] env[63088]: _type = "HttpNfcLease" [ 907.799735] env[63088]: } is ready. {{(pid=63088) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 907.800662] env[63088]: DEBUG oslo_vmware.rw_handles [None req-b81753c4-9726-4ca0-bfa8-881e9fef1e6f tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 907.800662] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52301661-3b30-c310-240c-271a50b2dabe" [ 907.800662] env[63088]: _type = "HttpNfcLease" [ 907.800662] env[63088]: }. {{(pid=63088) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 907.801511] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01d2889f-678b-4d7a-be2f-775a30231bd3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.811499] env[63088]: DEBUG oslo_vmware.rw_handles [None req-b81753c4-9726-4ca0-bfa8-881e9fef1e6f tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5220ae8f-5ee9-b64a-1a41-52ce2e6a51ec/disk-0.vmdk from lease info. {{(pid=63088) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 907.811727] env[63088]: DEBUG oslo_vmware.rw_handles [None req-b81753c4-9726-4ca0-bfa8-881e9fef1e6f tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5220ae8f-5ee9-b64a-1a41-52ce2e6a51ec/disk-0.vmdk for reading. {{(pid=63088) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 907.910365] env[63088]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-35fef36b-521a-4a83-860f-97307cf5b7e9 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.018838] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d90a9c5b-32f9-4a01-9f9b-3303d7e8794d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.026009] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f864096-94eb-4091-b50c-a387aa90401a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.056677] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a26c3c9-c92e-4850-8080-bced387e12a9 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.064469] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57ccb72d-43aa-4b44-a610-b63bcf3e52af {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.077569] env[63088]: DEBUG nova.compute.provider_tree [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 908.091243] env[63088]: DEBUG oslo_vmware.api [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': task-1285153, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.605461} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.092104] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] bebc3318-24bb-4a37-8b23-66a12a7f7fd2/bebc3318-24bb-4a37-8b23-66a12a7f7fd2.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 908.092354] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: bebc3318-24bb-4a37-8b23-66a12a7f7fd2] Extending root virtual disk to 1048576 {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 908.092595] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e5c5a650-e414-4e57-9904-7c27cfa4f0fd {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.098407] env[63088]: DEBUG oslo_vmware.api [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Waiting for the task: (returnval){ [ 908.098407] env[63088]: value = "task-1285157" [ 908.098407] env[63088]: _type = "Task" [ 908.098407] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.106372] env[63088]: DEBUG oslo_vmware.api [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': task-1285157, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.120098] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285156, 'name': CreateVM_Task, 'duration_secs': 0.442101} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.120298] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 45513c6b-c000-4ee1-8893-4e084ffc22c5] Created VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 908.120966] env[63088]: DEBUG oslo_concurrency.lockutils [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 908.121155] env[63088]: DEBUG oslo_concurrency.lockutils [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 908.121463] env[63088]: DEBUG oslo_concurrency.lockutils [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 908.121742] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-69cbe06d-0fe7-4417-aa58-d9ada06e0298 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.126268] env[63088]: DEBUG oslo_vmware.api [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Waiting for the task: (returnval){ [ 908.126268] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52133c65-c4e0-425c-3b6c-c591fa657067" [ 908.126268] env[63088]: _type = "Task" [ 908.126268] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.136894] env[63088]: DEBUG oslo_vmware.api [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52133c65-c4e0-425c-3b6c-c591fa657067, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.249735] env[63088]: DEBUG oslo_concurrency.lockutils [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquiring lock "fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.250035] env[63088]: DEBUG oslo_concurrency.lockutils [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lock "fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.296735] env[63088]: INFO nova.compute.manager [-] [instance: db032da3-d334-486f-a428-af1c8a3c360a] Took 1.48 seconds to deallocate network for instance. [ 908.332330] env[63088]: DEBUG nova.network.neutron [req-eef8ce2c-114a-401e-a4a3-a4405a370a29 req-45026a6b-b78b-4b4d-abe7-d52eb5454b3f service nova] [instance: 45513c6b-c000-4ee1-8893-4e084ffc22c5] Updated VIF entry in instance network info cache for port 3d848d39-f98f-4614-918d-aba9f290becb. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 908.332637] env[63088]: DEBUG nova.network.neutron [req-eef8ce2c-114a-401e-a4a3-a4405a370a29 req-45026a6b-b78b-4b4d-abe7-d52eb5454b3f service nova] [instance: 45513c6b-c000-4ee1-8893-4e084ffc22c5] Updating instance_info_cache with network_info: [{"id": "3d848d39-f98f-4614-918d-aba9f290becb", "address": "fa:16:3e:08:54:08", "network": {"id": "79155637-c844-4012-a77a-d04f9ec9ab18", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-2144942540-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e27be174c0b3400193f9ae64a242da1a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d848d39-f9", "ovs_interfaceid": "3d848d39-f98f-4614-918d-aba9f290becb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 908.581336] env[63088]: DEBUG nova.scheduler.client.report [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 908.612065] env[63088]: DEBUG oslo_vmware.api [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': task-1285157, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072065} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.612450] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: bebc3318-24bb-4a37-8b23-66a12a7f7fd2] Extended root virtual disk {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 908.613583] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9d0c138-063a-44f4-8e80-6c9c9da2ce93 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.637387] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: bebc3318-24bb-4a37-8b23-66a12a7f7fd2] Reconfiguring VM instance instance-0000004e to attach disk [datastore1] bebc3318-24bb-4a37-8b23-66a12a7f7fd2/bebc3318-24bb-4a37-8b23-66a12a7f7fd2.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 908.641176] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3bb50261-c03d-4828-aa07-f12be04ef24c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.661916] env[63088]: DEBUG oslo_vmware.api [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52133c65-c4e0-425c-3b6c-c591fa657067, 'name': SearchDatastore_Task, 'duration_secs': 0.011893} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.664358] env[63088]: DEBUG oslo_concurrency.lockutils [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 908.664693] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 45513c6b-c000-4ee1-8893-4e084ffc22c5] Processing image 1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 908.665026] env[63088]: DEBUG oslo_concurrency.lockutils [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 908.665236] env[63088]: DEBUG oslo_concurrency.lockutils [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 908.665452] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 908.665847] env[63088]: DEBUG oslo_vmware.api [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Waiting for the task: (returnval){ [ 908.665847] env[63088]: value = "task-1285158" [ 908.665847] env[63088]: _type = "Task" [ 908.665847] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.666101] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8f01eb60-0fa1-40e3-bee3-e7e0c1b8c4c3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.679382] env[63088]: DEBUG oslo_vmware.api [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': task-1285158, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.680903] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 908.681239] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 908.682098] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-19e2f4dd-71a5-404c-a060-443f4620fd0f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.688534] env[63088]: DEBUG oslo_vmware.api [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Waiting for the task: (returnval){ [ 908.688534] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]526ae242-8408-df9f-46d9-479cd163cafe" [ 908.688534] env[63088]: _type = "Task" [ 908.688534] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.697315] env[63088]: DEBUG oslo_vmware.api [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]526ae242-8408-df9f-46d9-479cd163cafe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.725086] env[63088]: DEBUG nova.objects.instance [None req-e2dee0e1-16ef-4ce0-9a3b-bc5d82868451 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Lazy-loading 'flavor' on Instance uuid dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 908.752846] env[63088]: DEBUG nova.compute.manager [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 908.805072] env[63088]: DEBUG oslo_concurrency.lockutils [None req-2bccf470-f018-4694-b367-d6d0046e5e91 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.834936] env[63088]: DEBUG oslo_concurrency.lockutils [req-eef8ce2c-114a-401e-a4a3-a4405a370a29 req-45026a6b-b78b-4b4d-abe7-d52eb5454b3f service nova] Releasing lock "refresh_cache-45513c6b-c000-4ee1-8893-4e084ffc22c5" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 909.088981] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63088) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 909.088981] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.402s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 909.088981] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c799bf18-c9a4-4b67-989b-6dda4cd4483c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.104s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 909.088981] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c799bf18-c9a4-4b67-989b-6dda4cd4483c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 909.093388] env[63088]: DEBUG oslo_concurrency.lockutils [None req-125c7f05-a006-44cd-8334-62558a6d840d tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.479s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 909.093613] env[63088]: DEBUG oslo_concurrency.lockutils [None req-125c7f05-a006-44cd-8334-62558a6d840d tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 909.095498] env[63088]: DEBUG oslo_concurrency.lockutils [None req-33fca862-7795-4260-b5ce-68be35c68550 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.007s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 909.095757] env[63088]: DEBUG oslo_concurrency.lockutils [None req-33fca862-7795-4260-b5ce-68be35c68550 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 909.097690] env[63088]: DEBUG oslo_concurrency.lockutils [None req-1bc44084-4c8b-4bbf-b317-1818ae32cad5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 2.679s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 909.125103] env[63088]: INFO nova.scheduler.client.report [None req-125c7f05-a006-44cd-8334-62558a6d840d tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Deleted allocations for instance 4971b24c-6710-4f50-9846-727dad264b1f [ 909.129846] env[63088]: INFO nova.scheduler.client.report [None req-33fca862-7795-4260-b5ce-68be35c68550 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Deleted allocations for instance 667fd5e9-5fe4-41e1-9d8a-896c6e5c6286 [ 909.138870] env[63088]: INFO nova.scheduler.client.report [None req-c799bf18-c9a4-4b67-989b-6dda4cd4483c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Deleted allocations for instance 9d5f1cde-e787-4a2c-ac49-83ec135ff51c [ 909.178918] env[63088]: DEBUG oslo_vmware.api [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': task-1285158, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.201611] env[63088]: DEBUG oslo_vmware.api [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]526ae242-8408-df9f-46d9-479cd163cafe, 'name': SearchDatastore_Task, 'duration_secs': 0.015504} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.203201] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-24bbd3bd-ff9c-49ee-af5b-688d98960f23 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.209539] env[63088]: DEBUG oslo_vmware.api [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Waiting for the task: (returnval){ [ 909.209539] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52164e3b-bbce-fd34-374e-7cddda7bccf7" [ 909.209539] env[63088]: _type = "Task" [ 909.209539] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.219067] env[63088]: DEBUG oslo_vmware.api [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52164e3b-bbce-fd34-374e-7cddda7bccf7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.230758] env[63088]: DEBUG oslo_concurrency.lockutils [None req-e2dee0e1-16ef-4ce0-9a3b-bc5d82868451 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Lock "dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.286s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 909.264835] env[63088]: DEBUG nova.compute.manager [req-4d5cc978-693d-4871-9a20-2b0abf18f5ce req-a15b5a3b-4e3f-40de-88d6-c71b1007716a service nova] [instance: db032da3-d334-486f-a428-af1c8a3c360a] Received event network-vif-deleted-98219984-3198-4e54-8c93-024446f958a3 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 909.265069] env[63088]: DEBUG nova.compute.manager [req-4d5cc978-693d-4871-9a20-2b0abf18f5ce req-a15b5a3b-4e3f-40de-88d6-c71b1007716a service nova] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Received event network-changed-79ca8279-a3e2-4792-a4b5-57f899be5c41 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 909.265255] env[63088]: DEBUG nova.compute.manager [req-4d5cc978-693d-4871-9a20-2b0abf18f5ce req-a15b5a3b-4e3f-40de-88d6-c71b1007716a service nova] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Refreshing instance network info cache due to event network-changed-79ca8279-a3e2-4792-a4b5-57f899be5c41. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 909.265480] env[63088]: DEBUG oslo_concurrency.lockutils [req-4d5cc978-693d-4871-9a20-2b0abf18f5ce req-a15b5a3b-4e3f-40de-88d6-c71b1007716a service nova] Acquiring lock "refresh_cache-71e37d8e-a454-46c4-a3cc-3d5671a32beb" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 909.265628] env[63088]: DEBUG oslo_concurrency.lockutils [req-4d5cc978-693d-4871-9a20-2b0abf18f5ce req-a15b5a3b-4e3f-40de-88d6-c71b1007716a service nova] Acquired lock "refresh_cache-71e37d8e-a454-46c4-a3cc-3d5671a32beb" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 909.265788] env[63088]: DEBUG nova.network.neutron [req-4d5cc978-693d-4871-9a20-2b0abf18f5ce req-a15b5a3b-4e3f-40de-88d6-c71b1007716a service nova] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Refreshing network info cache for port 79ca8279-a3e2-4792-a4b5-57f899be5c41 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 909.277671] env[63088]: DEBUG oslo_concurrency.lockutils [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 909.553070] env[63088]: DEBUG oslo_concurrency.lockutils [None req-197c58f9-6d72-4bbb-82c6-88b551caa18d tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquiring lock "2269e520-08b6-433f-8fe9-7b84d91e02d7" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 909.553070] env[63088]: DEBUG oslo_concurrency.lockutils [None req-197c58f9-6d72-4bbb-82c6-88b551caa18d tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Lock "2269e520-08b6-433f-8fe9-7b84d91e02d7" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 909.553070] env[63088]: DEBUG nova.compute.manager [None req-197c58f9-6d72-4bbb-82c6-88b551caa18d tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Going to confirm migration 2 {{(pid=63088) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 909.601447] env[63088]: DEBUG nova.objects.instance [None req-1bc44084-4c8b-4bbf-b317-1818ae32cad5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Lazy-loading 'migration_context' on Instance uuid e6b0ce08-d67f-458f-92f5-1e904d03a5ef {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 909.640948] env[63088]: DEBUG oslo_concurrency.lockutils [None req-125c7f05-a006-44cd-8334-62558a6d840d tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Lock "4971b24c-6710-4f50-9846-727dad264b1f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.361s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 909.642273] env[63088]: DEBUG oslo_concurrency.lockutils [req-ecb54d06-7451-4f48-9fcf-32c633776cb3 req-6d9c3d35-0d97-4c30-a811-7315cafc0f3a service nova] Acquired lock "4971b24c-6710-4f50-9846-727dad264b1f" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 909.644236] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b39f775-df87-46e5-a756-f6e94e4ee9c1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.652511] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c799bf18-c9a4-4b67-989b-6dda4cd4483c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "9d5f1cde-e787-4a2c-ac49-83ec135ff51c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.412s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 909.658203] env[63088]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 909.658438] env[63088]: DEBUG oslo_vmware.api [-] Fault list: [ManagedObjectNotFound] {{(pid=63088) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 909.660264] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5495b94c-d327-457f-85f9-a3edc81988e2 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.669167] env[63088]: DEBUG oslo_concurrency.lockutils [None req-33fca862-7795-4260-b5ce-68be35c68550 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Lock "667fd5e9-5fe4-41e1-9d8a-896c6e5c6286" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.883s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 909.674381] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f727cb7-5d4e-4d29-bebb-1fc8180a7e32 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.702767] env[63088]: DEBUG oslo_vmware.api [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': task-1285158, 'name': ReconfigVM_Task, 'duration_secs': 0.52547} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.703162] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: bebc3318-24bb-4a37-8b23-66a12a7f7fd2] Reconfigured VM instance instance-0000004e to attach disk [datastore1] bebc3318-24bb-4a37-8b23-66a12a7f7fd2/bebc3318-24bb-4a37-8b23-66a12a7f7fd2.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 909.704065] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e2a3b4e9-fa39-46a4-9271-0ed29ea292f8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.722390] env[63088]: ERROR root [req-ecb54d06-7451-4f48-9fcf-32c633776cb3 req-6d9c3d35-0d97-4c30-a811-7315cafc0f3a service nova] Original exception being dropped: ['Traceback (most recent call last):\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py", line 377, in request_handler\n response = request(managed_object, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 586, in __call__\n return client.invoke(args, kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 728, in invoke\n result = self.send(soapenv, timeout=timeout)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 777, in send\n return self.process_reply(reply.message, None, None)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 840, in process_reply\n raise WebFault(fault, replyroot)\n', "suds.WebFault: Server raised fault: 'The object 'vim.VirtualMachine:vm-275841' has already been deleted or has not been completely created'\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 301, in _invoke_api\n return api_method(*args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/vim_util.py", line 480, in get_object_property\n props = get_object_properties(vim, moref, [property_name],\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/vim_util.py", line 360, in get_object_properties\n retrieve_result = vim.RetrievePropertiesEx(\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py", line 413, in request_handler\n raise exceptions.VimFaultException(fault_list, fault_string,\n', "oslo_vmware.exceptions.VimFaultException: The object 'vim.VirtualMachine:vm-275841' has already been deleted or has not been completely created\nCause: Server raised fault: 'The object 'vim.VirtualMachine:vm-275841' has already been deleted or has not been completely created'\nFaults: [ManagedObjectNotFound]\nDetails: {'obj': 'vm-275841'}\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/session.py", line 123, in _call_method\n return self.invoke_api(module, method, self.vim, *args,\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 358, in invoke_api\n return _invoke_api(module, method, *args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 122, in func\n return evt.wait()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait\n result = hub.switch()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch\n return self.greenlet.switch()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/common/loopingcall.py", line 122, in _inner\n idle = self.f(*self.args, **self.kw)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 96, in _func\n result = f(*args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 341, in _invoke_api\n raise clazz(str(excep),\n', "oslo_vmware.exceptions.ManagedObjectNotFoundException: The object 'vim.VirtualMachine:vm-275841' has already been deleted or has not been completely created\nCause: Server raised fault: 'The object 'vim.VirtualMachine:vm-275841' has already been deleted or has not been completely created'\nFaults: [ManagedObjectNotFound]\nDetails: {'obj': 'vm-275841'}\n"]: nova.exception.InstanceNotFound: Instance 4971b24c-6710-4f50-9846-727dad264b1f could not be found. [ 909.723606] env[63088]: DEBUG oslo_concurrency.lockutils [req-ecb54d06-7451-4f48-9fcf-32c633776cb3 req-6d9c3d35-0d97-4c30-a811-7315cafc0f3a service nova] Releasing lock "4971b24c-6710-4f50-9846-727dad264b1f" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 909.724857] env[63088]: DEBUG nova.compute.manager [req-ecb54d06-7451-4f48-9fcf-32c633776cb3 req-6d9c3d35-0d97-4c30-a811-7315cafc0f3a service nova] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Detach interface failed, port_id=81932288-bc32-47c4-a3b2-02597cb6bf81, reason: Instance 4971b24c-6710-4f50-9846-727dad264b1f could not be found. {{(pid=63088) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 909.725144] env[63088]: DEBUG nova.compute.manager [req-ecb54d06-7451-4f48-9fcf-32c633776cb3 req-6d9c3d35-0d97-4c30-a811-7315cafc0f3a service nova] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Received event network-vif-deleted-592cc24a-92a9-4bd3-bb9a-a0877ac0db04 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 909.725459] env[63088]: INFO nova.compute.manager [req-ecb54d06-7451-4f48-9fcf-32c633776cb3 req-6d9c3d35-0d97-4c30-a811-7315cafc0f3a service nova] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Neutron deleted interface 592cc24a-92a9-4bd3-bb9a-a0877ac0db04; detaching it from the instance and deleting it from the info cache [ 909.725647] env[63088]: DEBUG nova.network.neutron [req-ecb54d06-7451-4f48-9fcf-32c633776cb3 req-6d9c3d35-0d97-4c30-a811-7315cafc0f3a service nova] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Updating instance_info_cache with network_info: [{"id": "d5a10951-ae2e-48f5-b0ca-b1f144e5fe3a", "address": "fa:16:3e:95:96:0c", "network": {"id": "1174a284-d3ca-4f9e-aa81-13ee9a693e55", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1994276040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.166", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa3d24a1a6c0430985fd80365d986ee1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd5a10951-ae", "ovs_interfaceid": "d5a10951-ae2e-48f5-b0ca-b1f144e5fe3a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 909.733503] env[63088]: DEBUG oslo_vmware.api [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Waiting for the task: (returnval){ [ 909.733503] env[63088]: value = "task-1285159" [ 909.733503] env[63088]: _type = "Task" [ 909.733503] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.737519] env[63088]: DEBUG oslo_vmware.api [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52164e3b-bbce-fd34-374e-7cddda7bccf7, 'name': SearchDatastore_Task, 'duration_secs': 0.017256} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.741078] env[63088]: DEBUG oslo_concurrency.lockutils [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 909.741378] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore2] 45513c6b-c000-4ee1-8893-4e084ffc22c5/45513c6b-c000-4ee1-8893-4e084ffc22c5.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 909.742290] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-103ca7ed-aa7b-4012-a32a-81bf39c84530 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.750424] env[63088]: DEBUG oslo_vmware.api [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': task-1285159, 'name': Rename_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.751973] env[63088]: DEBUG oslo_vmware.api [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Waiting for the task: (returnval){ [ 909.751973] env[63088]: value = "task-1285160" [ 909.751973] env[63088]: _type = "Task" [ 909.751973] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.762740] env[63088]: DEBUG oslo_vmware.api [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': task-1285160, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.067124] env[63088]: DEBUG nova.network.neutron [req-4d5cc978-693d-4871-9a20-2b0abf18f5ce req-a15b5a3b-4e3f-40de-88d6-c71b1007716a service nova] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Updated VIF entry in instance network info cache for port 79ca8279-a3e2-4792-a4b5-57f899be5c41. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 910.067564] env[63088]: DEBUG nova.network.neutron [req-4d5cc978-693d-4871-9a20-2b0abf18f5ce req-a15b5a3b-4e3f-40de-88d6-c71b1007716a service nova] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Updating instance_info_cache with network_info: [{"id": "79ca8279-a3e2-4792-a4b5-57f899be5c41", "address": "fa:16:3e:53:f8:3c", "network": {"id": "cc2c0304-8906-443a-a76e-e7a32158346f", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1867758042-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.139", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b18ddddf2314d83addf550b8cb91977", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0f096917-a0cf-4add-a9d2-23ca1c723b3b", "external-id": "nsx-vlan-transportzone-894", "segmentation_id": 894, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap79ca8279-a3", "ovs_interfaceid": "79ca8279-a3e2-4792-a4b5-57f899be5c41", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 910.136994] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b3780f6f-9e55-4577-8cb1-842578e7b6a6 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Acquiring lock "dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 910.137273] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b3780f6f-9e55-4577-8cb1-842578e7b6a6 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Lock "dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 910.149103] env[63088]: DEBUG oslo_concurrency.lockutils [None req-197c58f9-6d72-4bbb-82c6-88b551caa18d tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquiring lock "refresh_cache-2269e520-08b6-433f-8fe9-7b84d91e02d7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 910.149322] env[63088]: DEBUG oslo_concurrency.lockutils [None req-197c58f9-6d72-4bbb-82c6-88b551caa18d tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquired lock "refresh_cache-2269e520-08b6-433f-8fe9-7b84d91e02d7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 910.149512] env[63088]: DEBUG nova.network.neutron [None req-197c58f9-6d72-4bbb-82c6-88b551caa18d tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 910.149715] env[63088]: DEBUG nova.objects.instance [None req-197c58f9-6d72-4bbb-82c6-88b551caa18d tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Lazy-loading 'info_cache' on Instance uuid 2269e520-08b6-433f-8fe9-7b84d91e02d7 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 910.231391] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2685b7d4-bc7d-4da2-8c32-f3c87c192114 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.248401] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50a45c4f-ed74-4b77-82db-dcc8cf60060d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.267357] env[63088]: DEBUG oslo_vmware.api [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': task-1285159, 'name': Rename_Task, 'duration_secs': 0.179753} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.273163] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: bebc3318-24bb-4a37-8b23-66a12a7f7fd2] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 910.273163] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-886f184c-b6b2-4deb-a116-7ea489f20e4f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.282401] env[63088]: DEBUG oslo_vmware.api [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': task-1285160, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.297904] env[63088]: DEBUG nova.compute.manager [req-ecb54d06-7451-4f48-9fcf-32c633776cb3 req-6d9c3d35-0d97-4c30-a811-7315cafc0f3a service nova] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Detach interface failed, port_id=592cc24a-92a9-4bd3-bb9a-a0877ac0db04, reason: Instance 4971b24c-6710-4f50-9846-727dad264b1f could not be found. {{(pid=63088) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 910.298870] env[63088]: DEBUG oslo_vmware.api [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Waiting for the task: (returnval){ [ 910.298870] env[63088]: value = "task-1285161" [ 910.298870] env[63088]: _type = "Task" [ 910.298870] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.308948] env[63088]: DEBUG oslo_vmware.api [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': task-1285161, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.374906] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b7ca8f0-2415-4f3b-a2fb-2baa7a0311dc {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.385736] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a25c308-52dd-4524-aa2b-1bd4bab01ff6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.420499] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dd4a76d-7406-47aa-95c7-5852b7b33a8f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.429798] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af95d66b-84c7-445d-8de1-d640118f8e33 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.446479] env[63088]: DEBUG nova.compute.provider_tree [None req-1bc44084-4c8b-4bbf-b317-1818ae32cad5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 910.573136] env[63088]: DEBUG oslo_concurrency.lockutils [req-4d5cc978-693d-4871-9a20-2b0abf18f5ce req-a15b5a3b-4e3f-40de-88d6-c71b1007716a service nova] Releasing lock "refresh_cache-71e37d8e-a454-46c4-a3cc-3d5671a32beb" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 910.640653] env[63088]: DEBUG nova.compute.utils [None req-b3780f6f-9e55-4577-8cb1-842578e7b6a6 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 910.777359] env[63088]: DEBUG oslo_vmware.api [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': task-1285160, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.627384} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.777839] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore2] 45513c6b-c000-4ee1-8893-4e084ffc22c5/45513c6b-c000-4ee1-8893-4e084ffc22c5.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 910.777839] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 45513c6b-c000-4ee1-8893-4e084ffc22c5] Extending root virtual disk to 1048576 {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 910.778116] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dc829c56-dfa4-4419-8599-9ffaff3b46e8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.784386] env[63088]: DEBUG oslo_vmware.api [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Waiting for the task: (returnval){ [ 910.784386] env[63088]: value = "task-1285162" [ 910.784386] env[63088]: _type = "Task" [ 910.784386] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.792506] env[63088]: DEBUG oslo_vmware.api [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': task-1285162, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.808226] env[63088]: DEBUG oslo_vmware.api [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': task-1285161, 'name': PowerOnVM_Task} progress is 78%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.949637] env[63088]: DEBUG nova.scheduler.client.report [None req-1bc44084-4c8b-4bbf-b317-1818ae32cad5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 911.144764] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b3780f6f-9e55-4577-8cb1-842578e7b6a6 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Lock "dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.297867] env[63088]: DEBUG oslo_vmware.api [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': task-1285162, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.116017} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.298135] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 45513c6b-c000-4ee1-8893-4e084ffc22c5] Extended root virtual disk {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 911.299082] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-899cfc35-5970-4dcc-a3a1-47cdaef80847 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.328261] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 45513c6b-c000-4ee1-8893-4e084ffc22c5] Reconfiguring VM instance instance-0000004f to attach disk [datastore2] 45513c6b-c000-4ee1-8893-4e084ffc22c5/45513c6b-c000-4ee1-8893-4e084ffc22c5.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 911.334514] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ceaa4778-f874-4370-8659-dbc3752c561c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.351038] env[63088]: DEBUG oslo_vmware.api [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': task-1285161, 'name': PowerOnVM_Task, 'duration_secs': 0.793555} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.351749] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: bebc3318-24bb-4a37-8b23-66a12a7f7fd2] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 911.352189] env[63088]: INFO nova.compute.manager [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: bebc3318-24bb-4a37-8b23-66a12a7f7fd2] Took 8.76 seconds to spawn the instance on the hypervisor. [ 911.352189] env[63088]: DEBUG nova.compute.manager [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: bebc3318-24bb-4a37-8b23-66a12a7f7fd2] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 911.353417] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d7a89de-c557-4744-9728-5e1788d5bc36 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.357977] env[63088]: DEBUG oslo_vmware.api [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Waiting for the task: (returnval){ [ 911.357977] env[63088]: value = "task-1285163" [ 911.357977] env[63088]: _type = "Task" [ 911.357977] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.372781] env[63088]: DEBUG oslo_vmware.api [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': task-1285163, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.451139] env[63088]: DEBUG oslo_concurrency.lockutils [None req-937b14db-2197-4ed2-abfe-5c8f35ed9c52 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquiring lock "7e9f26a9-c988-4229-ac02-2e56955a9b60" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.451395] env[63088]: DEBUG oslo_concurrency.lockutils [None req-937b14db-2197-4ed2-abfe-5c8f35ed9c52 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "7e9f26a9-c988-4229-ac02-2e56955a9b60" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.577716] env[63088]: DEBUG nova.network.neutron [None req-197c58f9-6d72-4bbb-82c6-88b551caa18d tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Updating instance_info_cache with network_info: [{"id": "1fb0e170-03ab-4ee4-9f1f-c5c6008c823d", "address": "fa:16:3e:a8:02:36", "network": {"id": "dff14a7f-0af1-4e4e-a498-86d7c9816e6b", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-762073671-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f33f2701fad94864a8c406a404bc0a42", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "21310d90-efbc-45a8-a97f-c4358606530f", "external-id": "nsx-vlan-transportzone-672", "segmentation_id": 672, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1fb0e170-03", "ovs_interfaceid": "1fb0e170-03ab-4ee4-9f1f-c5c6008c823d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 911.880094] env[63088]: INFO nova.compute.manager [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: bebc3318-24bb-4a37-8b23-66a12a7f7fd2] Took 32.15 seconds to build instance. [ 911.885125] env[63088]: DEBUG oslo_vmware.api [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': task-1285163, 'name': ReconfigVM_Task, 'duration_secs': 0.412758} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.885885] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 45513c6b-c000-4ee1-8893-4e084ffc22c5] Reconfigured VM instance instance-0000004f to attach disk [datastore2] 45513c6b-c000-4ee1-8893-4e084ffc22c5/45513c6b-c000-4ee1-8893-4e084ffc22c5.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 911.886744] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-82fc6be6-a0f9-4ea2-a56b-aa92f5e41ae4 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.894664] env[63088]: DEBUG oslo_vmware.api [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Waiting for the task: (returnval){ [ 911.894664] env[63088]: value = "task-1285164" [ 911.894664] env[63088]: _type = "Task" [ 911.894664] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.907737] env[63088]: DEBUG oslo_vmware.api [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': task-1285164, 'name': Rename_Task} progress is 6%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.961143] env[63088]: DEBUG nova.compute.manager [None req-937b14db-2197-4ed2-abfe-5c8f35ed9c52 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 7e9f26a9-c988-4229-ac02-2e56955a9b60] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 911.965897] env[63088]: DEBUG oslo_concurrency.lockutils [None req-1bc44084-4c8b-4bbf-b317-1818ae32cad5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.868s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.972449] env[63088]: DEBUG oslo_concurrency.lockutils [None req-2bccf470-f018-4694-b367-d6d0046e5e91 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.168s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.972709] env[63088]: DEBUG nova.objects.instance [None req-2bccf470-f018-4694-b367-d6d0046e5e91 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Lazy-loading 'resources' on Instance uuid db032da3-d334-486f-a428-af1c8a3c360a {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 912.082789] env[63088]: DEBUG oslo_concurrency.lockutils [None req-197c58f9-6d72-4bbb-82c6-88b551caa18d tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Releasing lock "refresh_cache-2269e520-08b6-433f-8fe9-7b84d91e02d7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 912.083237] env[63088]: DEBUG nova.objects.instance [None req-197c58f9-6d72-4bbb-82c6-88b551caa18d tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Lazy-loading 'migration_context' on Instance uuid 2269e520-08b6-433f-8fe9-7b84d91e02d7 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 912.226455] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b3780f6f-9e55-4577-8cb1-842578e7b6a6 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Acquiring lock "dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 912.226850] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b3780f6f-9e55-4577-8cb1-842578e7b6a6 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Lock "dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 912.227164] env[63088]: INFO nova.compute.manager [None req-b3780f6f-9e55-4577-8cb1-842578e7b6a6 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Attaching volume 51f9c2dc-ac18-40cf-87db-5359fcb203a8 to /dev/sdc [ 912.266625] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3551692e-13f9-481b-8ec3-5b13bc0d4c61 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.274643] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2fad26f-3b5a-4ba8-b249-60ad69034ee3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.288089] env[63088]: DEBUG nova.virt.block_device [None req-b3780f6f-9e55-4577-8cb1-842578e7b6a6 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Updating existing volume attachment record: de0d9565-ee94-4666-8940-a5359aedcb48 {{(pid=63088) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 912.385517] env[63088]: DEBUG oslo_concurrency.lockutils [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Lock "bebc3318-24bb-4a37-8b23-66a12a7f7fd2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 33.660s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 912.408069] env[63088]: DEBUG oslo_vmware.api [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': task-1285164, 'name': Rename_Task, 'duration_secs': 0.170485} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.408482] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 45513c6b-c000-4ee1-8893-4e084ffc22c5] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 912.408831] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3aa09581-a51e-4671-a12b-e99685272996 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.416349] env[63088]: DEBUG oslo_vmware.api [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Waiting for the task: (returnval){ [ 912.416349] env[63088]: value = "task-1285165" [ 912.416349] env[63088]: _type = "Task" [ 912.416349] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.425359] env[63088]: DEBUG oslo_vmware.api [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': task-1285165, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.473407] env[63088]: DEBUG nova.compute.manager [None req-937b14db-2197-4ed2-abfe-5c8f35ed9c52 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 7e9f26a9-c988-4229-ac02-2e56955a9b60] Instance disappeared before build. {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 912.592035] env[63088]: DEBUG nova.objects.base [None req-197c58f9-6d72-4bbb-82c6-88b551caa18d tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Object Instance<2269e520-08b6-433f-8fe9-7b84d91e02d7> lazy-loaded attributes: info_cache,migration_context {{(pid=63088) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 912.594154] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-704b55a1-7a54-4165-a697-d5002e2cde15 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.618198] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7bd4c686-47eb-4dbe-baff-5ee528f0dd4e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.623800] env[63088]: DEBUG oslo_vmware.api [None req-197c58f9-6d72-4bbb-82c6-88b551caa18d tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Waiting for the task: (returnval){ [ 912.623800] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52fd9685-0c70-6312-59a8-102b1e7edc0d" [ 912.623800] env[63088]: _type = "Task" [ 912.623800] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.638791] env[63088]: DEBUG oslo_vmware.api [None req-197c58f9-6d72-4bbb-82c6-88b551caa18d tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52fd9685-0c70-6312-59a8-102b1e7edc0d, 'name': SearchDatastore_Task, 'duration_secs': 0.008182} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.639196] env[63088]: DEBUG oslo_concurrency.lockutils [None req-197c58f9-6d72-4bbb-82c6-88b551caa18d tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 912.737910] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf7ac0a7-11ae-42e9-9a84-87310c352d9f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.748077] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7f258bf-8d79-41b6-93af-bcd20a2b2f5f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.782044] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44661a1f-27c4-49da-8232-18f94920f8de {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.790282] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3a3d188-75fe-474d-b890-29eb1755bdd8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.805871] env[63088]: DEBUG nova.compute.provider_tree [None req-2bccf470-f018-4694-b367-d6d0046e5e91 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 912.930212] env[63088]: DEBUG oslo_vmware.api [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': task-1285165, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.991809] env[63088]: DEBUG oslo_concurrency.lockutils [None req-937b14db-2197-4ed2-abfe-5c8f35ed9c52 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "7e9f26a9-c988-4229-ac02-2e56955a9b60" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 1.540s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.031409] env[63088]: DEBUG oslo_concurrency.lockutils [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquiring lock "72c07d1e-cbb6-4875-9b7f-2649aa25eee5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.031738] env[63088]: DEBUG oslo_concurrency.lockutils [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "72c07d1e-cbb6-4875-9b7f-2649aa25eee5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.070276] env[63088]: DEBUG oslo_concurrency.lockutils [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquiring lock "f582da15-dfc3-45e1-a995-9dd0c9533869" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.070569] env[63088]: DEBUG oslo_concurrency.lockutils [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Lock "f582da15-dfc3-45e1-a995-9dd0c9533869" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.165835] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6545a09f-2d02-4a01-bf68-ba1a748f4bee tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Acquiring lock "b50db882-598c-488f-b935-34f55f655642" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.166066] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6545a09f-2d02-4a01-bf68-ba1a748f4bee tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Lock "b50db882-598c-488f-b935-34f55f655642" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.166294] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6545a09f-2d02-4a01-bf68-ba1a748f4bee tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Acquiring lock "b50db882-598c-488f-b935-34f55f655642-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.166483] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6545a09f-2d02-4a01-bf68-ba1a748f4bee tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Lock "b50db882-598c-488f-b935-34f55f655642-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.166658] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6545a09f-2d02-4a01-bf68-ba1a748f4bee tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Lock "b50db882-598c-488f-b935-34f55f655642-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.169040] env[63088]: INFO nova.compute.manager [None req-6545a09f-2d02-4a01-bf68-ba1a748f4bee tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: b50db882-598c-488f-b935-34f55f655642] Terminating instance [ 913.171201] env[63088]: DEBUG nova.compute.manager [None req-6545a09f-2d02-4a01-bf68-ba1a748f4bee tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: b50db882-598c-488f-b935-34f55f655642] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 913.171460] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-6545a09f-2d02-4a01-bf68-ba1a748f4bee tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: b50db882-598c-488f-b935-34f55f655642] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 913.172536] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cbcd754-aa16-4b44-b1a9-782a68166c4d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.181312] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-6545a09f-2d02-4a01-bf68-ba1a748f4bee tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: b50db882-598c-488f-b935-34f55f655642] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 913.181683] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bb834c76-0929-4164-b87a-2932e2715acf {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.189425] env[63088]: DEBUG oslo_vmware.api [None req-6545a09f-2d02-4a01-bf68-ba1a748f4bee tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Waiting for the task: (returnval){ [ 913.189425] env[63088]: value = "task-1285167" [ 913.189425] env[63088]: _type = "Task" [ 913.189425] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.197877] env[63088]: DEBUG oslo_vmware.api [None req-6545a09f-2d02-4a01-bf68-ba1a748f4bee tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Task: {'id': task-1285167, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.241982] env[63088]: DEBUG oslo_concurrency.lockutils [None req-cd65c449-e671-45aa-a201-5e68b37dd782 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Acquiring lock "2964db35-7357-40a7-b4e6-7e2595549f5b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.242377] env[63088]: DEBUG oslo_concurrency.lockutils [None req-cd65c449-e671-45aa-a201-5e68b37dd782 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Lock "2964db35-7357-40a7-b4e6-7e2595549f5b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.242678] env[63088]: DEBUG oslo_concurrency.lockutils [None req-cd65c449-e671-45aa-a201-5e68b37dd782 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Acquiring lock "2964db35-7357-40a7-b4e6-7e2595549f5b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.242948] env[63088]: DEBUG oslo_concurrency.lockutils [None req-cd65c449-e671-45aa-a201-5e68b37dd782 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Lock "2964db35-7357-40a7-b4e6-7e2595549f5b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.243152] env[63088]: DEBUG oslo_concurrency.lockutils [None req-cd65c449-e671-45aa-a201-5e68b37dd782 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Lock "2964db35-7357-40a7-b4e6-7e2595549f5b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.245558] env[63088]: INFO nova.compute.manager [None req-cd65c449-e671-45aa-a201-5e68b37dd782 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: 2964db35-7357-40a7-b4e6-7e2595549f5b] Terminating instance [ 913.247820] env[63088]: DEBUG nova.compute.manager [None req-cd65c449-e671-45aa-a201-5e68b37dd782 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: 2964db35-7357-40a7-b4e6-7e2595549f5b] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 913.248169] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-cd65c449-e671-45aa-a201-5e68b37dd782 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: 2964db35-7357-40a7-b4e6-7e2595549f5b] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 913.249092] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c916e025-ab16-445d-821c-5b6fdf003b96 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.258322] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd65c449-e671-45aa-a201-5e68b37dd782 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: 2964db35-7357-40a7-b4e6-7e2595549f5b] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 913.258689] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e44a72c4-3ccc-46e6-ae12-6b694469993e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.265609] env[63088]: DEBUG oslo_vmware.api [None req-cd65c449-e671-45aa-a201-5e68b37dd782 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Waiting for the task: (returnval){ [ 913.265609] env[63088]: value = "task-1285168" [ 913.265609] env[63088]: _type = "Task" [ 913.265609] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.276276] env[63088]: DEBUG oslo_vmware.api [None req-cd65c449-e671-45aa-a201-5e68b37dd782 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Task: {'id': task-1285168, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.308905] env[63088]: DEBUG nova.scheduler.client.report [None req-2bccf470-f018-4694-b367-d6d0046e5e91 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 913.428479] env[63088]: DEBUG oslo_vmware.api [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': task-1285165, 'name': PowerOnVM_Task, 'duration_secs': 0.590928} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.428868] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 45513c6b-c000-4ee1-8893-4e084ffc22c5] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 913.429155] env[63088]: INFO nova.compute.manager [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 45513c6b-c000-4ee1-8893-4e084ffc22c5] Took 8.42 seconds to spawn the instance on the hypervisor. [ 913.429403] env[63088]: DEBUG nova.compute.manager [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 45513c6b-c000-4ee1-8893-4e084ffc22c5] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 913.430416] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19324e18-4ffc-4487-8cda-39fcd66ecd56 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.511024] env[63088]: INFO nova.compute.manager [None req-1bc44084-4c8b-4bbf-b317-1818ae32cad5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Swapping old allocation on dict_keys(['6eae54a9-8831-40eb-bf54-4bc60d346b02']) held by migration 02e19d97-e38c-43e6-b757-056d827d87dc for instance [ 913.534594] env[63088]: DEBUG nova.scheduler.client.report [None req-1bc44084-4c8b-4bbf-b317-1818ae32cad5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Overwriting current allocation {'allocations': {'6eae54a9-8831-40eb-bf54-4bc60d346b02': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 96}}, 'project_id': '982309216b714182b5304c4d15808b9e', 'user_id': '9b77ca5186e5437a862d644b2075fe8b', 'consumer_generation': 1} on consumer e6b0ce08-d67f-458f-92f5-1e904d03a5ef {{(pid=63088) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 913.536724] env[63088]: DEBUG nova.compute.manager [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 913.573071] env[63088]: DEBUG nova.compute.manager [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 913.622677] env[63088]: DEBUG oslo_concurrency.lockutils [None req-1bc44084-4c8b-4bbf-b317-1818ae32cad5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Acquiring lock "refresh_cache-e6b0ce08-d67f-458f-92f5-1e904d03a5ef" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 913.622904] env[63088]: DEBUG oslo_concurrency.lockutils [None req-1bc44084-4c8b-4bbf-b317-1818ae32cad5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Acquired lock "refresh_cache-e6b0ce08-d67f-458f-92f5-1e904d03a5ef" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 913.623702] env[63088]: DEBUG nova.network.neutron [None req-1bc44084-4c8b-4bbf-b317-1818ae32cad5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 913.699846] env[63088]: DEBUG oslo_vmware.api [None req-6545a09f-2d02-4a01-bf68-ba1a748f4bee tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Task: {'id': task-1285167, 'name': PowerOffVM_Task, 'duration_secs': 0.240558} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.700193] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-6545a09f-2d02-4a01-bf68-ba1a748f4bee tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: b50db882-598c-488f-b935-34f55f655642] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 913.700378] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-6545a09f-2d02-4a01-bf68-ba1a748f4bee tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: b50db882-598c-488f-b935-34f55f655642] Unregistering the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 913.700679] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e6aa25dd-7334-4a8b-a185-9f1a454d9c82 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.765708] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-6545a09f-2d02-4a01-bf68-ba1a748f4bee tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: b50db882-598c-488f-b935-34f55f655642] Unregistered the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 913.765708] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-6545a09f-2d02-4a01-bf68-ba1a748f4bee tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: b50db882-598c-488f-b935-34f55f655642] Deleting contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 913.765708] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-6545a09f-2d02-4a01-bf68-ba1a748f4bee tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Deleting the datastore file [datastore1] b50db882-598c-488f-b935-34f55f655642 {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 913.765708] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-63a4c1b3-d6d6-4b1b-a978-84040dbf93ba {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.776169] env[63088]: DEBUG oslo_vmware.api [None req-cd65c449-e671-45aa-a201-5e68b37dd782 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Task: {'id': task-1285168, 'name': PowerOffVM_Task, 'duration_secs': 0.293564} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.777466] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd65c449-e671-45aa-a201-5e68b37dd782 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: 2964db35-7357-40a7-b4e6-7e2595549f5b] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 913.777652] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-cd65c449-e671-45aa-a201-5e68b37dd782 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: 2964db35-7357-40a7-b4e6-7e2595549f5b] Unregistering the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 913.778015] env[63088]: DEBUG oslo_vmware.api [None req-6545a09f-2d02-4a01-bf68-ba1a748f4bee tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Waiting for the task: (returnval){ [ 913.778015] env[63088]: value = "task-1285170" [ 913.778015] env[63088]: _type = "Task" [ 913.778015] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.778240] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b5978794-edc6-410a-9d7d-4c978dad19de {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.787203] env[63088]: DEBUG oslo_vmware.api [None req-6545a09f-2d02-4a01-bf68-ba1a748f4bee tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Task: {'id': task-1285170, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.814430] env[63088]: DEBUG oslo_concurrency.lockutils [None req-2bccf470-f018-4694-b367-d6d0046e5e91 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.842s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.816926] env[63088]: DEBUG oslo_concurrency.lockutils [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.539s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.818687] env[63088]: INFO nova.compute.claims [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 913.833890] env[63088]: INFO nova.scheduler.client.report [None req-2bccf470-f018-4694-b367-d6d0046e5e91 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Deleted allocations for instance db032da3-d334-486f-a428-af1c8a3c360a [ 913.843860] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-cd65c449-e671-45aa-a201-5e68b37dd782 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: 2964db35-7357-40a7-b4e6-7e2595549f5b] Unregistered the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 913.844117] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-cd65c449-e671-45aa-a201-5e68b37dd782 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: 2964db35-7357-40a7-b4e6-7e2595549f5b] Deleting contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 913.844309] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-cd65c449-e671-45aa-a201-5e68b37dd782 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Deleting the datastore file [datastore1] 2964db35-7357-40a7-b4e6-7e2595549f5b {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 913.844590] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-791d3c53-2ca1-4d08-88a1-dd56509fb987 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.851541] env[63088]: DEBUG oslo_vmware.api [None req-cd65c449-e671-45aa-a201-5e68b37dd782 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Waiting for the task: (returnval){ [ 913.851541] env[63088]: value = "task-1285172" [ 913.851541] env[63088]: _type = "Task" [ 913.851541] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.859825] env[63088]: DEBUG oslo_vmware.api [None req-cd65c449-e671-45aa-a201-5e68b37dd782 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Task: {'id': task-1285172, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.956686] env[63088]: INFO nova.compute.manager [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 45513c6b-c000-4ee1-8893-4e084ffc22c5] Took 34.17 seconds to build instance. [ 914.097490] env[63088]: DEBUG oslo_concurrency.lockutils [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 914.098782] env[63088]: DEBUG oslo_concurrency.lockutils [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 914.296803] env[63088]: DEBUG oslo_vmware.api [None req-6545a09f-2d02-4a01-bf68-ba1a748f4bee tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Task: {'id': task-1285170, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.226252} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.297146] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-6545a09f-2d02-4a01-bf68-ba1a748f4bee tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 914.297388] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-6545a09f-2d02-4a01-bf68-ba1a748f4bee tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: b50db882-598c-488f-b935-34f55f655642] Deleted contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 914.297619] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-6545a09f-2d02-4a01-bf68-ba1a748f4bee tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: b50db882-598c-488f-b935-34f55f655642] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 914.297871] env[63088]: INFO nova.compute.manager [None req-6545a09f-2d02-4a01-bf68-ba1a748f4bee tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: b50db882-598c-488f-b935-34f55f655642] Took 1.13 seconds to destroy the instance on the hypervisor. [ 914.298205] env[63088]: DEBUG oslo.service.loopingcall [None req-6545a09f-2d02-4a01-bf68-ba1a748f4bee tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 914.298452] env[63088]: DEBUG nova.compute.manager [-] [instance: b50db882-598c-488f-b935-34f55f655642] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 914.298566] env[63088]: DEBUG nova.network.neutron [-] [instance: b50db882-598c-488f-b935-34f55f655642] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 914.343300] env[63088]: DEBUG oslo_concurrency.lockutils [None req-2bccf470-f018-4694-b367-d6d0046e5e91 tempest-ServerRescueTestJSON-546384677 tempest-ServerRescueTestJSON-546384677-project-member] Lock "db032da3-d334-486f-a428-af1c8a3c360a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.696s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.364107] env[63088]: DEBUG oslo_vmware.api [None req-cd65c449-e671-45aa-a201-5e68b37dd782 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Task: {'id': task-1285172, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.230146} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.364527] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-cd65c449-e671-45aa-a201-5e68b37dd782 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 914.364743] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-cd65c449-e671-45aa-a201-5e68b37dd782 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: 2964db35-7357-40a7-b4e6-7e2595549f5b] Deleted contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 914.364927] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-cd65c449-e671-45aa-a201-5e68b37dd782 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: 2964db35-7357-40a7-b4e6-7e2595549f5b] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 914.365128] env[63088]: INFO nova.compute.manager [None req-cd65c449-e671-45aa-a201-5e68b37dd782 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] [instance: 2964db35-7357-40a7-b4e6-7e2595549f5b] Took 1.12 seconds to destroy the instance on the hypervisor. [ 914.365378] env[63088]: DEBUG oslo.service.loopingcall [None req-cd65c449-e671-45aa-a201-5e68b37dd782 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 914.366256] env[63088]: DEBUG nova.compute.manager [-] [instance: 2964db35-7357-40a7-b4e6-7e2595549f5b] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 914.366355] env[63088]: DEBUG nova.network.neutron [-] [instance: 2964db35-7357-40a7-b4e6-7e2595549f5b] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 914.459196] env[63088]: DEBUG oslo_concurrency.lockutils [None req-61fa3e8f-5876-4dce-8696-3ae06233d6de tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Lock "45513c6b-c000-4ee1-8893-4e084ffc22c5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.682s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.465657] env[63088]: DEBUG nova.network.neutron [None req-1bc44084-4c8b-4bbf-b317-1818ae32cad5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Updating instance_info_cache with network_info: [{"id": "63d8dbbe-4550-4d8d-8e85-9a9eadc81d82", "address": "fa:16:3e:d1:72:56", "network": {"id": "ebeefbc2-043c-4e76-aeaf-4f9113869855", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.60", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "bfb6bc5062fc4d51af1d3c577659be9a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271fe7a0-dfd7-409b-920a-cf83ef1a86a3", "external-id": "nsx-vlan-transportzone-728", "segmentation_id": 728, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap63d8dbbe-45", "ovs_interfaceid": "63d8dbbe-4550-4d8d-8e85-9a9eadc81d82", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 914.503544] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4acba9f7-d735-4a08-90c3-58649f241640 tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Acquiring lock "bebc3318-24bb-4a37-8b23-66a12a7f7fd2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 914.503544] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4acba9f7-d735-4a08-90c3-58649f241640 tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Lock "bebc3318-24bb-4a37-8b23-66a12a7f7fd2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.503544] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4acba9f7-d735-4a08-90c3-58649f241640 tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Acquiring lock "bebc3318-24bb-4a37-8b23-66a12a7f7fd2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 914.503544] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4acba9f7-d735-4a08-90c3-58649f241640 tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Lock "bebc3318-24bb-4a37-8b23-66a12a7f7fd2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.503802] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4acba9f7-d735-4a08-90c3-58649f241640 tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Lock "bebc3318-24bb-4a37-8b23-66a12a7f7fd2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.506953] env[63088]: INFO nova.compute.manager [None req-4acba9f7-d735-4a08-90c3-58649f241640 tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: bebc3318-24bb-4a37-8b23-66a12a7f7fd2] Terminating instance [ 914.514224] env[63088]: DEBUG nova.compute.manager [None req-4acba9f7-d735-4a08-90c3-58649f241640 tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: bebc3318-24bb-4a37-8b23-66a12a7f7fd2] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 914.514224] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-4acba9f7-d735-4a08-90c3-58649f241640 tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: bebc3318-24bb-4a37-8b23-66a12a7f7fd2] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 914.514224] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e31556f4-03be-407b-8b95-fc2f2def5195 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.522793] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-4acba9f7-d735-4a08-90c3-58649f241640 tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: bebc3318-24bb-4a37-8b23-66a12a7f7fd2] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 914.523452] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f3b2a420-ce64-4486-aa46-a89868994a72 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.530557] env[63088]: DEBUG oslo_vmware.api [None req-4acba9f7-d735-4a08-90c3-58649f241640 tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Waiting for the task: (returnval){ [ 914.530557] env[63088]: value = "task-1285173" [ 914.530557] env[63088]: _type = "Task" [ 914.530557] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.539625] env[63088]: DEBUG oslo_vmware.api [None req-4acba9f7-d735-4a08-90c3-58649f241640 tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': task-1285173, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.607192] env[63088]: DEBUG oslo_concurrency.lockutils [None req-901dfe38-2626-4a52-ab00-0a33f4bc1b1f tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Acquiring lock "45513c6b-c000-4ee1-8893-4e084ffc22c5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 914.607382] env[63088]: DEBUG oslo_concurrency.lockutils [None req-901dfe38-2626-4a52-ab00-0a33f4bc1b1f tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Lock "45513c6b-c000-4ee1-8893-4e084ffc22c5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.607609] env[63088]: DEBUG oslo_concurrency.lockutils [None req-901dfe38-2626-4a52-ab00-0a33f4bc1b1f tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Acquiring lock "45513c6b-c000-4ee1-8893-4e084ffc22c5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 914.607795] env[63088]: DEBUG oslo_concurrency.lockutils [None req-901dfe38-2626-4a52-ab00-0a33f4bc1b1f tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Lock "45513c6b-c000-4ee1-8893-4e084ffc22c5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.608020] env[63088]: DEBUG oslo_concurrency.lockutils [None req-901dfe38-2626-4a52-ab00-0a33f4bc1b1f tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Lock "45513c6b-c000-4ee1-8893-4e084ffc22c5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.612181] env[63088]: INFO nova.compute.manager [None req-901dfe38-2626-4a52-ab00-0a33f4bc1b1f tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 45513c6b-c000-4ee1-8893-4e084ffc22c5] Terminating instance [ 914.617303] env[63088]: DEBUG nova.compute.manager [None req-901dfe38-2626-4a52-ab00-0a33f4bc1b1f tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 45513c6b-c000-4ee1-8893-4e084ffc22c5] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 914.618578] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-901dfe38-2626-4a52-ab00-0a33f4bc1b1f tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 45513c6b-c000-4ee1-8893-4e084ffc22c5] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 914.619662] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c269bdfe-5871-4172-9c79-aa756ab552b2 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.631076] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-901dfe38-2626-4a52-ab00-0a33f4bc1b1f tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 45513c6b-c000-4ee1-8893-4e084ffc22c5] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 914.631379] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9f9ed7bb-d4ae-4a99-9c9b-d331feb36d0c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.638080] env[63088]: DEBUG oslo_vmware.api [None req-901dfe38-2626-4a52-ab00-0a33f4bc1b1f tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Waiting for the task: (returnval){ [ 914.638080] env[63088]: value = "task-1285175" [ 914.638080] env[63088]: _type = "Task" [ 914.638080] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.648775] env[63088]: DEBUG oslo_vmware.api [None req-901dfe38-2626-4a52-ab00-0a33f4bc1b1f tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': task-1285175, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.651591] env[63088]: DEBUG nova.compute.manager [req-dd5920f7-b879-4509-a391-8bc601a54644 req-e26a9b3a-f59a-4355-93b6-fb06ec1f8434 service nova] [instance: b50db882-598c-488f-b935-34f55f655642] Received event network-vif-deleted-585051d5-6974-4abd-864e-38967371f048 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 914.651924] env[63088]: INFO nova.compute.manager [req-dd5920f7-b879-4509-a391-8bc601a54644 req-e26a9b3a-f59a-4355-93b6-fb06ec1f8434 service nova] [instance: b50db882-598c-488f-b935-34f55f655642] Neutron deleted interface 585051d5-6974-4abd-864e-38967371f048; detaching it from the instance and deleting it from the info cache [ 914.652017] env[63088]: DEBUG nova.network.neutron [req-dd5920f7-b879-4509-a391-8bc601a54644 req-e26a9b3a-f59a-4355-93b6-fb06ec1f8434 service nova] [instance: b50db882-598c-488f-b935-34f55f655642] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 914.672484] env[63088]: DEBUG nova.compute.manager [req-f796a7cb-dd51-415e-8289-61f364aa650f req-1bb8f952-798a-41e6-9041-318b77c0d949 service nova] [instance: 2964db35-7357-40a7-b4e6-7e2595549f5b] Received event network-vif-deleted-b42d0eed-48b4-4937-b222-80a994ea7602 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 914.672667] env[63088]: INFO nova.compute.manager [req-f796a7cb-dd51-415e-8289-61f364aa650f req-1bb8f952-798a-41e6-9041-318b77c0d949 service nova] [instance: 2964db35-7357-40a7-b4e6-7e2595549f5b] Neutron deleted interface b42d0eed-48b4-4937-b222-80a994ea7602; detaching it from the instance and deleting it from the info cache [ 914.672846] env[63088]: DEBUG nova.network.neutron [req-f796a7cb-dd51-415e-8289-61f364aa650f req-1bb8f952-798a-41e6-9041-318b77c0d949 service nova] [instance: 2964db35-7357-40a7-b4e6-7e2595549f5b] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 914.968499] env[63088]: DEBUG oslo_concurrency.lockutils [None req-1bc44084-4c8b-4bbf-b317-1818ae32cad5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Releasing lock "refresh_cache-e6b0ce08-d67f-458f-92f5-1e904d03a5ef" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.968974] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-1bc44084-4c8b-4bbf-b317-1818ae32cad5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 914.969350] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d376cb68-d93f-44ca-9f45-a193224a3982 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.977293] env[63088]: DEBUG oslo_vmware.api [None req-1bc44084-4c8b-4bbf-b317-1818ae32cad5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Waiting for the task: (returnval){ [ 914.977293] env[63088]: value = "task-1285176" [ 914.977293] env[63088]: _type = "Task" [ 914.977293] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.989525] env[63088]: DEBUG oslo_vmware.api [None req-1bc44084-4c8b-4bbf-b317-1818ae32cad5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Task: {'id': task-1285176, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.045059] env[63088]: DEBUG oslo_vmware.api [None req-4acba9f7-d735-4a08-90c3-58649f241640 tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': task-1285173, 'name': PowerOffVM_Task, 'duration_secs': 0.273281} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.045401] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-4acba9f7-d735-4a08-90c3-58649f241640 tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: bebc3318-24bb-4a37-8b23-66a12a7f7fd2] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 915.045583] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-4acba9f7-d735-4a08-90c3-58649f241640 tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: bebc3318-24bb-4a37-8b23-66a12a7f7fd2] Unregistering the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 915.045889] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c07a93aa-d47f-4992-ab7d-718c2906d5c9 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.064490] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a57323c-e34a-4efa-810c-e19d17e68057 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.072980] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61092906-3792-482f-a8e6-de2ade5c6d92 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.107598] env[63088]: DEBUG nova.network.neutron [-] [instance: b50db882-598c-488f-b935-34f55f655642] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 915.111918] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc4669fb-1465-4b18-90d8-33b544778710 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.115086] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-4acba9f7-d735-4a08-90c3-58649f241640 tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: bebc3318-24bb-4a37-8b23-66a12a7f7fd2] Unregistered the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 915.115328] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-4acba9f7-d735-4a08-90c3-58649f241640 tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: bebc3318-24bb-4a37-8b23-66a12a7f7fd2] Deleting contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 915.115535] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-4acba9f7-d735-4a08-90c3-58649f241640 tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Deleting the datastore file [datastore1] bebc3318-24bb-4a37-8b23-66a12a7f7fd2 {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 915.115858] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a432fda4-cf48-48cc-a441-04590045e852 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.124702] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b974acf2-8a45-499d-a39e-780c4cd7af61 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.128842] env[63088]: DEBUG oslo_vmware.api [None req-4acba9f7-d735-4a08-90c3-58649f241640 tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Waiting for the task: (returnval){ [ 915.128842] env[63088]: value = "task-1285178" [ 915.128842] env[63088]: _type = "Task" [ 915.128842] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.142573] env[63088]: DEBUG nova.compute.provider_tree [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 915.149542] env[63088]: DEBUG nova.network.neutron [-] [instance: 2964db35-7357-40a7-b4e6-7e2595549f5b] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 915.155532] env[63088]: DEBUG oslo_vmware.api [None req-4acba9f7-d735-4a08-90c3-58649f241640 tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': task-1285178, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.156277] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-15fe82ef-3301-4a64-998d-edb29be2bdb5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.164547] env[63088]: DEBUG oslo_vmware.api [None req-901dfe38-2626-4a52-ab00-0a33f4bc1b1f tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': task-1285175, 'name': PowerOffVM_Task, 'duration_secs': 0.294335} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.165388] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-901dfe38-2626-4a52-ab00-0a33f4bc1b1f tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 45513c6b-c000-4ee1-8893-4e084ffc22c5] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 915.165619] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-901dfe38-2626-4a52-ab00-0a33f4bc1b1f tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 45513c6b-c000-4ee1-8893-4e084ffc22c5] Unregistering the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 915.166314] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d5509c34-ce6f-4011-b5a6-9fb2f14fb389 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.172908] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66f75103-1e31-4039-b12c-4cb65a7456b3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.185490] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e1054f0f-7abd-43bc-9413-37dacbba1e23 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.194090] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f731120b-b92a-4542-8327-bfcd8b14a06b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.214389] env[63088]: DEBUG nova.compute.manager [req-dd5920f7-b879-4509-a391-8bc601a54644 req-e26a9b3a-f59a-4355-93b6-fb06ec1f8434 service nova] [instance: b50db882-598c-488f-b935-34f55f655642] Detach interface failed, port_id=585051d5-6974-4abd-864e-38967371f048, reason: Instance b50db882-598c-488f-b935-34f55f655642 could not be found. {{(pid=63088) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 915.232679] env[63088]: DEBUG nova.compute.manager [req-f796a7cb-dd51-415e-8289-61f364aa650f req-1bb8f952-798a-41e6-9041-318b77c0d949 service nova] [instance: 2964db35-7357-40a7-b4e6-7e2595549f5b] Detach interface failed, port_id=b42d0eed-48b4-4937-b222-80a994ea7602, reason: Instance 2964db35-7357-40a7-b4e6-7e2595549f5b could not be found. {{(pid=63088) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 915.244892] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-901dfe38-2626-4a52-ab00-0a33f4bc1b1f tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 45513c6b-c000-4ee1-8893-4e084ffc22c5] Unregistered the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 915.245155] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-901dfe38-2626-4a52-ab00-0a33f4bc1b1f tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 45513c6b-c000-4ee1-8893-4e084ffc22c5] Deleting contents of the VM from datastore datastore2 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 915.245348] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-901dfe38-2626-4a52-ab00-0a33f4bc1b1f tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Deleting the datastore file [datastore2] 45513c6b-c000-4ee1-8893-4e084ffc22c5 {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 915.245622] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9b137f8b-22b6-41a5-8b32-8b276697f685 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.255023] env[63088]: DEBUG oslo_vmware.api [None req-901dfe38-2626-4a52-ab00-0a33f4bc1b1f tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Waiting for the task: (returnval){ [ 915.255023] env[63088]: value = "task-1285180" [ 915.255023] env[63088]: _type = "Task" [ 915.255023] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.259998] env[63088]: DEBUG oslo_vmware.api [None req-901dfe38-2626-4a52-ab00-0a33f4bc1b1f tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': task-1285180, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.486847] env[63088]: DEBUG oslo_vmware.api [None req-1bc44084-4c8b-4bbf-b317-1818ae32cad5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Task: {'id': task-1285176, 'name': PowerOffVM_Task, 'duration_secs': 0.238398} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.487121] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-1bc44084-4c8b-4bbf-b317-1818ae32cad5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 915.487837] env[63088]: DEBUG nova.virt.hardware [None req-1bc44084-4c8b-4bbf-b317-1818ae32cad5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:18:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='99d18326-0562-4fb9-afc4-e9e0d5c92bb4',id=38,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-1781853035',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 915.488106] env[63088]: DEBUG nova.virt.hardware [None req-1bc44084-4c8b-4bbf-b317-1818ae32cad5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 915.488290] env[63088]: DEBUG nova.virt.hardware [None req-1bc44084-4c8b-4bbf-b317-1818ae32cad5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 915.488486] env[63088]: DEBUG nova.virt.hardware [None req-1bc44084-4c8b-4bbf-b317-1818ae32cad5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 915.488771] env[63088]: DEBUG nova.virt.hardware [None req-1bc44084-4c8b-4bbf-b317-1818ae32cad5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 915.488867] env[63088]: DEBUG nova.virt.hardware [None req-1bc44084-4c8b-4bbf-b317-1818ae32cad5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 915.489028] env[63088]: DEBUG nova.virt.hardware [None req-1bc44084-4c8b-4bbf-b317-1818ae32cad5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 915.489217] env[63088]: DEBUG nova.virt.hardware [None req-1bc44084-4c8b-4bbf-b317-1818ae32cad5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 915.489399] env[63088]: DEBUG nova.virt.hardware [None req-1bc44084-4c8b-4bbf-b317-1818ae32cad5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 915.489579] env[63088]: DEBUG nova.virt.hardware [None req-1bc44084-4c8b-4bbf-b317-1818ae32cad5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 915.489768] env[63088]: DEBUG nova.virt.hardware [None req-1bc44084-4c8b-4bbf-b317-1818ae32cad5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 915.494881] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-76776679-aa55-43b5-9144-521662c8e4d3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.510363] env[63088]: DEBUG oslo_vmware.api [None req-1bc44084-4c8b-4bbf-b317-1818ae32cad5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Waiting for the task: (returnval){ [ 915.510363] env[63088]: value = "task-1285181" [ 915.510363] env[63088]: _type = "Task" [ 915.510363] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.519457] env[63088]: DEBUG oslo_vmware.api [None req-1bc44084-4c8b-4bbf-b317-1818ae32cad5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Task: {'id': task-1285181, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.612257] env[63088]: INFO nova.compute.manager [-] [instance: b50db882-598c-488f-b935-34f55f655642] Took 1.31 seconds to deallocate network for instance. [ 915.640711] env[63088]: DEBUG oslo_vmware.api [None req-4acba9f7-d735-4a08-90c3-58649f241640 tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': task-1285178, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.268471} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.641795] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-4acba9f7-d735-4a08-90c3-58649f241640 tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 915.641795] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-4acba9f7-d735-4a08-90c3-58649f241640 tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: bebc3318-24bb-4a37-8b23-66a12a7f7fd2] Deleted contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 915.641795] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-4acba9f7-d735-4a08-90c3-58649f241640 tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: bebc3318-24bb-4a37-8b23-66a12a7f7fd2] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 915.641955] env[63088]: INFO nova.compute.manager [None req-4acba9f7-d735-4a08-90c3-58649f241640 tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: bebc3318-24bb-4a37-8b23-66a12a7f7fd2] Took 1.13 seconds to destroy the instance on the hypervisor. [ 915.642323] env[63088]: DEBUG oslo.service.loopingcall [None req-4acba9f7-d735-4a08-90c3-58649f241640 tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 915.642598] env[63088]: DEBUG nova.compute.manager [-] [instance: bebc3318-24bb-4a37-8b23-66a12a7f7fd2] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 915.642700] env[63088]: DEBUG nova.network.neutron [-] [instance: bebc3318-24bb-4a37-8b23-66a12a7f7fd2] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 915.645412] env[63088]: DEBUG nova.scheduler.client.report [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 915.651590] env[63088]: INFO nova.compute.manager [-] [instance: 2964db35-7357-40a7-b4e6-7e2595549f5b] Took 1.29 seconds to deallocate network for instance. [ 915.762693] env[63088]: DEBUG oslo_vmware.api [None req-901dfe38-2626-4a52-ab00-0a33f4bc1b1f tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': task-1285180, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.020632] env[63088]: DEBUG oslo_vmware.api [None req-1bc44084-4c8b-4bbf-b317-1818ae32cad5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Task: {'id': task-1285181, 'name': ReconfigVM_Task, 'duration_secs': 0.237537} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.021504] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f7270ad-85f9-47b0-a548-20e820473f6f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.039640] env[63088]: DEBUG nova.virt.hardware [None req-1bc44084-4c8b-4bbf-b317-1818ae32cad5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:18:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='99d18326-0562-4fb9-afc4-e9e0d5c92bb4',id=38,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-1781853035',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 916.039898] env[63088]: DEBUG nova.virt.hardware [None req-1bc44084-4c8b-4bbf-b317-1818ae32cad5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 916.040096] env[63088]: DEBUG nova.virt.hardware [None req-1bc44084-4c8b-4bbf-b317-1818ae32cad5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 916.040297] env[63088]: DEBUG nova.virt.hardware [None req-1bc44084-4c8b-4bbf-b317-1818ae32cad5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 916.040448] env[63088]: DEBUG nova.virt.hardware [None req-1bc44084-4c8b-4bbf-b317-1818ae32cad5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 916.040597] env[63088]: DEBUG nova.virt.hardware [None req-1bc44084-4c8b-4bbf-b317-1818ae32cad5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 916.040798] env[63088]: DEBUG nova.virt.hardware [None req-1bc44084-4c8b-4bbf-b317-1818ae32cad5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 916.040962] env[63088]: DEBUG nova.virt.hardware [None req-1bc44084-4c8b-4bbf-b317-1818ae32cad5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 916.041188] env[63088]: DEBUG nova.virt.hardware [None req-1bc44084-4c8b-4bbf-b317-1818ae32cad5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 916.041375] env[63088]: DEBUG nova.virt.hardware [None req-1bc44084-4c8b-4bbf-b317-1818ae32cad5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 916.041553] env[63088]: DEBUG nova.virt.hardware [None req-1bc44084-4c8b-4bbf-b317-1818ae32cad5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 916.042393] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-279bed81-5a63-407d-bffa-2aea01e8b1e7 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.047570] env[63088]: DEBUG oslo_vmware.api [None req-1bc44084-4c8b-4bbf-b317-1818ae32cad5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Waiting for the task: (returnval){ [ 916.047570] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]5236f008-60b9-d6f2-2079-f1547f3107ab" [ 916.047570] env[63088]: _type = "Task" [ 916.047570] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.057156] env[63088]: DEBUG oslo_vmware.api [None req-1bc44084-4c8b-4bbf-b317-1818ae32cad5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]5236f008-60b9-d6f2-2079-f1547f3107ab, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.120450] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6545a09f-2d02-4a01-bf68-ba1a748f4bee tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.151817] env[63088]: DEBUG oslo_concurrency.lockutils [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.335s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 916.152436] env[63088]: DEBUG nova.compute.manager [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 916.155286] env[63088]: DEBUG oslo_concurrency.lockutils [None req-197c58f9-6d72-4bbb-82c6-88b551caa18d tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 3.516s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 916.163868] env[63088]: DEBUG oslo_concurrency.lockutils [None req-cd65c449-e671-45aa-a201-5e68b37dd782 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.261972] env[63088]: DEBUG oslo_vmware.api [None req-901dfe38-2626-4a52-ab00-0a33f4bc1b1f tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Task: {'id': task-1285180, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.915176} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.262263] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-901dfe38-2626-4a52-ab00-0a33f4bc1b1f tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 916.262787] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-901dfe38-2626-4a52-ab00-0a33f4bc1b1f tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 45513c6b-c000-4ee1-8893-4e084ffc22c5] Deleted contents of the VM from datastore datastore2 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 916.262787] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-901dfe38-2626-4a52-ab00-0a33f4bc1b1f tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 45513c6b-c000-4ee1-8893-4e084ffc22c5] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 916.262787] env[63088]: INFO nova.compute.manager [None req-901dfe38-2626-4a52-ab00-0a33f4bc1b1f tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] [instance: 45513c6b-c000-4ee1-8893-4e084ffc22c5] Took 1.65 seconds to destroy the instance on the hypervisor. [ 916.263042] env[63088]: DEBUG oslo.service.loopingcall [None req-901dfe38-2626-4a52-ab00-0a33f4bc1b1f tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 916.263525] env[63088]: DEBUG nova.compute.manager [-] [instance: 45513c6b-c000-4ee1-8893-4e084ffc22c5] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 916.263525] env[63088]: DEBUG nova.network.neutron [-] [instance: 45513c6b-c000-4ee1-8893-4e084ffc22c5] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 916.384920] env[63088]: DEBUG nova.network.neutron [-] [instance: bebc3318-24bb-4a37-8b23-66a12a7f7fd2] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 916.557368] env[63088]: DEBUG oslo_vmware.api [None req-1bc44084-4c8b-4bbf-b317-1818ae32cad5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]5236f008-60b9-d6f2-2079-f1547f3107ab, 'name': SearchDatastore_Task, 'duration_secs': 0.009487} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.563116] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-1bc44084-4c8b-4bbf-b317-1818ae32cad5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Reconfiguring VM instance instance-00000044 to detach disk 2000 {{(pid=63088) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 916.563509] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-def9cdf0-5f2e-4f74-92dc-4b7efcca9f79 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.585998] env[63088]: DEBUG oslo_vmware.api [None req-1bc44084-4c8b-4bbf-b317-1818ae32cad5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Waiting for the task: (returnval){ [ 916.585998] env[63088]: value = "task-1285182" [ 916.585998] env[63088]: _type = "Task" [ 916.585998] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.594892] env[63088]: DEBUG oslo_vmware.api [None req-1bc44084-4c8b-4bbf-b317-1818ae32cad5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Task: {'id': task-1285182, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.659057] env[63088]: DEBUG nova.compute.utils [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 916.664238] env[63088]: DEBUG nova.compute.manager [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 916.664495] env[63088]: DEBUG nova.network.neutron [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 916.720608] env[63088]: DEBUG nova.compute.manager [req-86cc00d1-d31d-4bbe-937b-a843fa84e853 req-39787eb7-6382-468a-836d-601423c17097 service nova] [instance: bebc3318-24bb-4a37-8b23-66a12a7f7fd2] Received event network-vif-deleted-bd817fe9-5d27-40bf-9348-e29243b263f5 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 916.720821] env[63088]: DEBUG nova.compute.manager [req-86cc00d1-d31d-4bbe-937b-a843fa84e853 req-39787eb7-6382-468a-836d-601423c17097 service nova] [instance: 45513c6b-c000-4ee1-8893-4e084ffc22c5] Received event network-vif-deleted-3d848d39-f98f-4614-918d-aba9f290becb {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 916.721444] env[63088]: INFO nova.compute.manager [req-86cc00d1-d31d-4bbe-937b-a843fa84e853 req-39787eb7-6382-468a-836d-601423c17097 service nova] [instance: 45513c6b-c000-4ee1-8893-4e084ffc22c5] Neutron deleted interface 3d848d39-f98f-4614-918d-aba9f290becb; detaching it from the instance and deleting it from the info cache [ 916.721444] env[63088]: DEBUG nova.network.neutron [req-86cc00d1-d31d-4bbe-937b-a843fa84e853 req-39787eb7-6382-468a-836d-601423c17097 service nova] [instance: 45513c6b-c000-4ee1-8893-4e084ffc22c5] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 916.736139] env[63088]: DEBUG nova.policy [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '32d5ff9b3ea345e4ac56010aa917e51b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '03e20d606e654362acbe2b36fe499ae3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 916.834880] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-b3780f6f-9e55-4577-8cb1-842578e7b6a6 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Volume attach. Driver type: vmdk {{(pid=63088) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 916.835182] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-b3780f6f-9e55-4577-8cb1-842578e7b6a6 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-275907', 'volume_id': '51f9c2dc-ac18-40cf-87db-5359fcb203a8', 'name': 'volume-51f9c2dc-ac18-40cf-87db-5359fcb203a8', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69', 'attached_at': '', 'detached_at': '', 'volume_id': '51f9c2dc-ac18-40cf-87db-5359fcb203a8', 'serial': '51f9c2dc-ac18-40cf-87db-5359fcb203a8'} {{(pid=63088) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 916.836212] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d711505d-6939-4d19-bdd3-ba9d0e01bd0c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.861198] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-deda02de-ab16-4110-8318-001af70bc8bf {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.893791] env[63088]: INFO nova.compute.manager [-] [instance: bebc3318-24bb-4a37-8b23-66a12a7f7fd2] Took 1.25 seconds to deallocate network for instance. [ 916.902610] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-b3780f6f-9e55-4577-8cb1-842578e7b6a6 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Reconfiguring VM instance instance-00000043 to attach disk [datastore1] volume-51f9c2dc-ac18-40cf-87db-5359fcb203a8/volume-51f9c2dc-ac18-40cf-87db-5359fcb203a8.vmdk or device None with type thin {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 916.909352] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f0406bff-1af2-4d83-ab45-9e5783c5a736 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.933845] env[63088]: DEBUG oslo_vmware.api [None req-b3780f6f-9e55-4577-8cb1-842578e7b6a6 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Waiting for the task: (returnval){ [ 916.933845] env[63088]: value = "task-1285183" [ 916.933845] env[63088]: _type = "Task" [ 916.933845] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.940088] env[63088]: DEBUG oslo_vmware.rw_handles [None req-b81753c4-9726-4ca0-bfa8-881e9fef1e6f tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5220ae8f-5ee9-b64a-1a41-52ce2e6a51ec/disk-0.vmdk. {{(pid=63088) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 916.941127] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdf01344-ff0f-421d-9e76-6b07a8c83c4a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.945501] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd029e4e-6580-4ee9-8ef0-4de8aff2fab0 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.953031] env[63088]: DEBUG oslo_vmware.api [None req-b3780f6f-9e55-4577-8cb1-842578e7b6a6 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Task: {'id': task-1285183, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.955757] env[63088]: DEBUG oslo_vmware.rw_handles [None req-b81753c4-9726-4ca0-bfa8-881e9fef1e6f tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5220ae8f-5ee9-b64a-1a41-52ce2e6a51ec/disk-0.vmdk is in state: ready. {{(pid=63088) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 916.956183] env[63088]: ERROR oslo_vmware.rw_handles [None req-b81753c4-9726-4ca0-bfa8-881e9fef1e6f tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5220ae8f-5ee9-b64a-1a41-52ce2e6a51ec/disk-0.vmdk due to incomplete transfer. [ 916.958762] env[63088]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-184079eb-4a1b-4a51-a4e8-e808396dc59e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.960996] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38c34915-ea28-4da1-8a2c-ff7879d51d83 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.998807] env[63088]: DEBUG nova.network.neutron [-] [instance: 45513c6b-c000-4ee1-8893-4e084ffc22c5] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 917.001995] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d5d4bf9-438a-40a2-b2bb-1bbb1b8bf00e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.004410] env[63088]: DEBUG oslo_vmware.rw_handles [None req-b81753c4-9726-4ca0-bfa8-881e9fef1e6f tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5220ae8f-5ee9-b64a-1a41-52ce2e6a51ec/disk-0.vmdk. {{(pid=63088) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 917.004642] env[63088]: DEBUG nova.virt.vmwareapi.images [None req-b81753c4-9726-4ca0-bfa8-881e9fef1e6f tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 254db932-35f1-42e3-9207-cd886efd65a3] Uploaded image c89225f6-2d34-4988-ae47-d3d6903890da to the Glance image server {{(pid=63088) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 917.006700] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-b81753c4-9726-4ca0-bfa8-881e9fef1e6f tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 254db932-35f1-42e3-9207-cd886efd65a3] Destroying the VM {{(pid=63088) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 917.007414] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-ec352766-e4c6-4c9c-817d-b048fc823ced {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.016881] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fd1c4f0-af7e-4d38-be7a-7edad0dc573d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.021262] env[63088]: DEBUG oslo_vmware.api [None req-b81753c4-9726-4ca0-bfa8-881e9fef1e6f tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Waiting for the task: (returnval){ [ 917.021262] env[63088]: value = "task-1285184" [ 917.021262] env[63088]: _type = "Task" [ 917.021262] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.032959] env[63088]: DEBUG nova.compute.provider_tree [None req-197c58f9-6d72-4bbb-82c6-88b551caa18d tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 917.038878] env[63088]: DEBUG oslo_vmware.api [None req-b81753c4-9726-4ca0-bfa8-881e9fef1e6f tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285184, 'name': Destroy_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.098677] env[63088]: DEBUG oslo_vmware.api [None req-1bc44084-4c8b-4bbf-b317-1818ae32cad5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Task: {'id': task-1285182, 'name': ReconfigVM_Task, 'duration_secs': 0.244819} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.098920] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-1bc44084-4c8b-4bbf-b317-1818ae32cad5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Reconfigured VM instance instance-00000044 to detach disk 2000 {{(pid=63088) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 917.100066] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a40379d3-54ba-4e6c-873a-82a2ea25b6fc {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.127970] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-1bc44084-4c8b-4bbf-b317-1818ae32cad5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Reconfiguring VM instance instance-00000044 to attach disk [datastore2] e6b0ce08-d67f-458f-92f5-1e904d03a5ef/e6b0ce08-d67f-458f-92f5-1e904d03a5ef.vmdk or device None with type thin {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 917.128370] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-14d56898-3573-4a03-bd0d-4351006643a0 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.147947] env[63088]: DEBUG oslo_vmware.api [None req-1bc44084-4c8b-4bbf-b317-1818ae32cad5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Waiting for the task: (returnval){ [ 917.147947] env[63088]: value = "task-1285185" [ 917.147947] env[63088]: _type = "Task" [ 917.147947] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.155699] env[63088]: DEBUG oslo_vmware.api [None req-1bc44084-4c8b-4bbf-b317-1818ae32cad5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Task: {'id': task-1285185, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.164666] env[63088]: DEBUG nova.compute.manager [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 917.224413] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0689fadc-4440-412f-bc43-2a1c25b7ffa4 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.233862] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22968bc9-04b8-4121-b85c-096c6183d401 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.261332] env[63088]: DEBUG nova.compute.manager [req-86cc00d1-d31d-4bbe-937b-a843fa84e853 req-39787eb7-6382-468a-836d-601423c17097 service nova] [instance: 45513c6b-c000-4ee1-8893-4e084ffc22c5] Detach interface failed, port_id=3d848d39-f98f-4614-918d-aba9f290becb, reason: Instance 45513c6b-c000-4ee1-8893-4e084ffc22c5 could not be found. {{(pid=63088) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 917.292758] env[63088]: DEBUG nova.network.neutron [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3] Successfully created port: 7e7a24aa-ad61-4e47-a533-6be5f25f7a37 {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 917.429875] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4acba9f7-d735-4a08-90c3-58649f241640 tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 917.442984] env[63088]: DEBUG oslo_vmware.api [None req-b3780f6f-9e55-4577-8cb1-842578e7b6a6 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Task: {'id': task-1285183, 'name': ReconfigVM_Task, 'duration_secs': 0.432538} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.443352] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-b3780f6f-9e55-4577-8cb1-842578e7b6a6 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Reconfigured VM instance instance-00000043 to attach disk [datastore1] volume-51f9c2dc-ac18-40cf-87db-5359fcb203a8/volume-51f9c2dc-ac18-40cf-87db-5359fcb203a8.vmdk or device None with type thin {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 917.448163] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-761f1ba8-82a1-4e7e-9263-c75586c9353f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.463673] env[63088]: DEBUG oslo_vmware.api [None req-b3780f6f-9e55-4577-8cb1-842578e7b6a6 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Waiting for the task: (returnval){ [ 917.463673] env[63088]: value = "task-1285186" [ 917.463673] env[63088]: _type = "Task" [ 917.463673] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.471999] env[63088]: DEBUG oslo_vmware.api [None req-b3780f6f-9e55-4577-8cb1-842578e7b6a6 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Task: {'id': task-1285186, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.507944] env[63088]: INFO nova.compute.manager [-] [instance: 45513c6b-c000-4ee1-8893-4e084ffc22c5] Took 1.24 seconds to deallocate network for instance. [ 917.530965] env[63088]: DEBUG oslo_vmware.api [None req-b81753c4-9726-4ca0-bfa8-881e9fef1e6f tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285184, 'name': Destroy_Task, 'duration_secs': 0.312034} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.531159] env[63088]: INFO nova.virt.vmwareapi.vm_util [None req-b81753c4-9726-4ca0-bfa8-881e9fef1e6f tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 254db932-35f1-42e3-9207-cd886efd65a3] Destroyed the VM [ 917.531482] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-b81753c4-9726-4ca0-bfa8-881e9fef1e6f tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 254db932-35f1-42e3-9207-cd886efd65a3] Deleting Snapshot of the VM instance {{(pid=63088) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 917.531733] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-524f3baf-c3f3-4208-bb29-c632bd902ef5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.537329] env[63088]: DEBUG oslo_vmware.api [None req-b81753c4-9726-4ca0-bfa8-881e9fef1e6f tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Waiting for the task: (returnval){ [ 917.537329] env[63088]: value = "task-1285187" [ 917.537329] env[63088]: _type = "Task" [ 917.537329] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.541768] env[63088]: DEBUG nova.scheduler.client.report [None req-197c58f9-6d72-4bbb-82c6-88b551caa18d tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 917.549718] env[63088]: DEBUG oslo_vmware.api [None req-b81753c4-9726-4ca0-bfa8-881e9fef1e6f tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285187, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.657946] env[63088]: DEBUG oslo_vmware.api [None req-1bc44084-4c8b-4bbf-b317-1818ae32cad5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Task: {'id': task-1285185, 'name': ReconfigVM_Task, 'duration_secs': 0.297336} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.658291] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-1bc44084-4c8b-4bbf-b317-1818ae32cad5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Reconfigured VM instance instance-00000044 to attach disk [datastore2] e6b0ce08-d67f-458f-92f5-1e904d03a5ef/e6b0ce08-d67f-458f-92f5-1e904d03a5ef.vmdk or device None with type thin {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 917.659140] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-866bb914-695f-431b-bc86-4a006fdb6909 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.681854] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cd75efc-4d5a-48cf-ae7c-e27dd4e2c213 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.700403] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97c4ef5f-e123-4a8d-b8e2-7e42a21c7f3c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.721394] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8ce6615-00dd-4304-a6d2-7d0d65007145 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.729217] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-1bc44084-4c8b-4bbf-b317-1818ae32cad5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 917.729501] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f814e7cc-c4b2-4402-b670-05a1e3a11a15 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.735647] env[63088]: DEBUG oslo_vmware.api [None req-1bc44084-4c8b-4bbf-b317-1818ae32cad5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Waiting for the task: (returnval){ [ 917.735647] env[63088]: value = "task-1285188" [ 917.735647] env[63088]: _type = "Task" [ 917.735647] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.743702] env[63088]: DEBUG oslo_vmware.api [None req-1bc44084-4c8b-4bbf-b317-1818ae32cad5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Task: {'id': task-1285188, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.973577] env[63088]: DEBUG oslo_vmware.api [None req-b3780f6f-9e55-4577-8cb1-842578e7b6a6 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Task: {'id': task-1285186, 'name': ReconfigVM_Task, 'duration_secs': 0.12936} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.973973] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-b3780f6f-9e55-4577-8cb1-842578e7b6a6 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-275907', 'volume_id': '51f9c2dc-ac18-40cf-87db-5359fcb203a8', 'name': 'volume-51f9c2dc-ac18-40cf-87db-5359fcb203a8', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69', 'attached_at': '', 'detached_at': '', 'volume_id': '51f9c2dc-ac18-40cf-87db-5359fcb203a8', 'serial': '51f9c2dc-ac18-40cf-87db-5359fcb203a8'} {{(pid=63088) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 918.014366] env[63088]: DEBUG oslo_concurrency.lockutils [None req-901dfe38-2626-4a52-ab00-0a33f4bc1b1f tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 918.052518] env[63088]: DEBUG oslo_vmware.api [None req-b81753c4-9726-4ca0-bfa8-881e9fef1e6f tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285187, 'name': RemoveSnapshot_Task} progress is 97%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.186646] env[63088]: DEBUG nova.compute.manager [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 918.209533] env[63088]: DEBUG nova.virt.hardware [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 918.209784] env[63088]: DEBUG nova.virt.hardware [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 918.209948] env[63088]: DEBUG nova.virt.hardware [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 918.210188] env[63088]: DEBUG nova.virt.hardware [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 918.210353] env[63088]: DEBUG nova.virt.hardware [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 918.210509] env[63088]: DEBUG nova.virt.hardware [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 918.210735] env[63088]: DEBUG nova.virt.hardware [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 918.210903] env[63088]: DEBUG nova.virt.hardware [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 918.211144] env[63088]: DEBUG nova.virt.hardware [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 918.211338] env[63088]: DEBUG nova.virt.hardware [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 918.211521] env[63088]: DEBUG nova.virt.hardware [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 918.212416] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3500164-1f27-40e4-a4d9-ac67a65b8186 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.220833] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a285a6b2-c7fe-4a34-87b0-fc204c191236 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.244524] env[63088]: DEBUG oslo_vmware.api [None req-1bc44084-4c8b-4bbf-b317-1818ae32cad5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Task: {'id': task-1285188, 'name': PowerOnVM_Task, 'duration_secs': 0.395551} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.244852] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-1bc44084-4c8b-4bbf-b317-1818ae32cad5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 918.549236] env[63088]: DEBUG oslo_vmware.api [None req-b81753c4-9726-4ca0-bfa8-881e9fef1e6f tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285187, 'name': RemoveSnapshot_Task, 'duration_secs': 0.771837} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.549538] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-b81753c4-9726-4ca0-bfa8-881e9fef1e6f tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 254db932-35f1-42e3-9207-cd886efd65a3] Deleted Snapshot of the VM instance {{(pid=63088) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 918.550150] env[63088]: INFO nova.compute.manager [None req-b81753c4-9726-4ca0-bfa8-881e9fef1e6f tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 254db932-35f1-42e3-9207-cd886efd65a3] Took 16.44 seconds to snapshot the instance on the hypervisor. [ 918.554014] env[63088]: DEBUG oslo_concurrency.lockutils [None req-197c58f9-6d72-4bbb-82c6-88b551caa18d tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.399s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.556736] env[63088]: DEBUG oslo_concurrency.lockutils [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.460s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 918.558122] env[63088]: INFO nova.compute.claims [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 918.748132] env[63088]: DEBUG nova.compute.manager [req-e48f01f4-2e3c-49ac-b7df-27caa0bf808a req-9d932f89-d14f-4595-bc5f-9eb5062a69d2 service nova] [instance: fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3] Received event network-vif-plugged-7e7a24aa-ad61-4e47-a533-6be5f25f7a37 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 918.748374] env[63088]: DEBUG oslo_concurrency.lockutils [req-e48f01f4-2e3c-49ac-b7df-27caa0bf808a req-9d932f89-d14f-4595-bc5f-9eb5062a69d2 service nova] Acquiring lock "fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 918.748583] env[63088]: DEBUG oslo_concurrency.lockutils [req-e48f01f4-2e3c-49ac-b7df-27caa0bf808a req-9d932f89-d14f-4595-bc5f-9eb5062a69d2 service nova] Lock "fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 918.748751] env[63088]: DEBUG oslo_concurrency.lockutils [req-e48f01f4-2e3c-49ac-b7df-27caa0bf808a req-9d932f89-d14f-4595-bc5f-9eb5062a69d2 service nova] Lock "fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.748921] env[63088]: DEBUG nova.compute.manager [req-e48f01f4-2e3c-49ac-b7df-27caa0bf808a req-9d932f89-d14f-4595-bc5f-9eb5062a69d2 service nova] [instance: fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3] No waiting events found dispatching network-vif-plugged-7e7a24aa-ad61-4e47-a533-6be5f25f7a37 {{(pid=63088) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 918.749295] env[63088]: WARNING nova.compute.manager [req-e48f01f4-2e3c-49ac-b7df-27caa0bf808a req-9d932f89-d14f-4595-bc5f-9eb5062a69d2 service nova] [instance: fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3] Received unexpected event network-vif-plugged-7e7a24aa-ad61-4e47-a533-6be5f25f7a37 for instance with vm_state building and task_state spawning. [ 918.764261] env[63088]: DEBUG nova.network.neutron [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3] Successfully updated port: 7e7a24aa-ad61-4e47-a533-6be5f25f7a37 {{(pid=63088) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 918.814565] env[63088]: DEBUG oslo_concurrency.lockutils [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Acquiring lock "79c4d6e8-9999-4b0c-98d3-bbfd4d61e660" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 918.814722] env[63088]: DEBUG oslo_concurrency.lockutils [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Lock "79c4d6e8-9999-4b0c-98d3-bbfd4d61e660" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 919.009027] env[63088]: DEBUG nova.objects.instance [None req-b3780f6f-9e55-4577-8cb1-842578e7b6a6 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Lazy-loading 'flavor' on Instance uuid dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 919.118052] env[63088]: INFO nova.scheduler.client.report [None req-197c58f9-6d72-4bbb-82c6-88b551caa18d tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Deleted allocation for migration 5602a4d9-9018-483a-98f7-e1e412ff902a [ 919.256026] env[63088]: INFO nova.compute.manager [None req-1bc44084-4c8b-4bbf-b317-1818ae32cad5 tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Updating instance to original state: 'active' [ 919.267393] env[63088]: DEBUG oslo_concurrency.lockutils [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquiring lock "refresh_cache-fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 919.267569] env[63088]: DEBUG oslo_concurrency.lockutils [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquired lock "refresh_cache-fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 919.271030] env[63088]: DEBUG nova.network.neutron [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 919.318012] env[63088]: DEBUG nova.compute.manager [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 79c4d6e8-9999-4b0c-98d3-bbfd4d61e660] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 919.515838] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b3780f6f-9e55-4577-8cb1-842578e7b6a6 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Lock "dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.289s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.624113] env[63088]: DEBUG oslo_concurrency.lockutils [None req-197c58f9-6d72-4bbb-82c6-88b551caa18d tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Lock "2269e520-08b6-433f-8fe9-7b84d91e02d7" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 10.071s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.751325] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-170638c6-15c7-4019-8f87-c6ad2119bb59 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.758971] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da147c78-716e-459e-8705-17d55b3a4648 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.794712] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f489e855-166c-43fc-9af5-4a7e4b88f9d7 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.803218] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e52706d-2468-4050-bd08-ebe9b04e6ecb {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.819438] env[63088]: DEBUG nova.compute.provider_tree [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 919.836508] env[63088]: DEBUG oslo_concurrency.lockutils [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 919.841711] env[63088]: DEBUG nova.network.neutron [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 919.961365] env[63088]: DEBUG oslo_concurrency.lockutils [None req-694d2129-1c60-4a36-8f0d-cb4e69a818c3 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Acquiring lock "dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 919.961824] env[63088]: DEBUG oslo_concurrency.lockutils [None req-694d2129-1c60-4a36-8f0d-cb4e69a818c3 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Lock "dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.009596] env[63088]: DEBUG nova.network.neutron [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3] Updating instance_info_cache with network_info: [{"id": "7e7a24aa-ad61-4e47-a533-6be5f25f7a37", "address": "fa:16:3e:f2:bc:c0", "network": {"id": "307b966a-d9e1-40fd-9313-1ad94c734308", "bridge": "br-int", "label": "tempest-ServersTestJSON-1907249197-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "03e20d606e654362acbe2b36fe499ae3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae18b41f-e73c-44f1-83dd-467c080944f4", "external-id": "nsx-vlan-transportzone-653", "segmentation_id": 653, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e7a24aa-ad", "ovs_interfaceid": "7e7a24aa-ad61-4e47-a533-6be5f25f7a37", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 920.322883] env[63088]: DEBUG nova.scheduler.client.report [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 920.466827] env[63088]: INFO nova.compute.manager [None req-694d2129-1c60-4a36-8f0d-cb4e69a818c3 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Detaching volume deedf893-eb6c-4ff0-ab0b-3e34abe88c8f [ 920.484844] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5a7bffe6-5d81-4eff-949f-17e8fd61a581 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquiring lock "2269e520-08b6-433f-8fe9-7b84d91e02d7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.484844] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5a7bffe6-5d81-4eff-949f-17e8fd61a581 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Lock "2269e520-08b6-433f-8fe9-7b84d91e02d7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.485122] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5a7bffe6-5d81-4eff-949f-17e8fd61a581 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquiring lock "2269e520-08b6-433f-8fe9-7b84d91e02d7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.485210] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5a7bffe6-5d81-4eff-949f-17e8fd61a581 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Lock "2269e520-08b6-433f-8fe9-7b84d91e02d7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.485410] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5a7bffe6-5d81-4eff-949f-17e8fd61a581 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Lock "2269e520-08b6-433f-8fe9-7b84d91e02d7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.488526] env[63088]: INFO nova.compute.manager [None req-5a7bffe6-5d81-4eff-949f-17e8fd61a581 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Terminating instance [ 920.493289] env[63088]: DEBUG nova.compute.manager [None req-5a7bffe6-5d81-4eff-949f-17e8fd61a581 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 920.493501] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-5a7bffe6-5d81-4eff-949f-17e8fd61a581 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 920.494440] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50e44ef0-de0a-4be6-9846-746caa3a37cc {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.503640] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a7bffe6-5d81-4eff-949f-17e8fd61a581 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 920.503958] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-19a16df9-17b7-4149-b48f-f5ab94cadd6c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.512164] env[63088]: DEBUG oslo_concurrency.lockutils [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Releasing lock "refresh_cache-fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 920.512572] env[63088]: DEBUG nova.compute.manager [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3] Instance network_info: |[{"id": "7e7a24aa-ad61-4e47-a533-6be5f25f7a37", "address": "fa:16:3e:f2:bc:c0", "network": {"id": "307b966a-d9e1-40fd-9313-1ad94c734308", "bridge": "br-int", "label": "tempest-ServersTestJSON-1907249197-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "03e20d606e654362acbe2b36fe499ae3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae18b41f-e73c-44f1-83dd-467c080944f4", "external-id": "nsx-vlan-transportzone-653", "segmentation_id": 653, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e7a24aa-ad", "ovs_interfaceid": "7e7a24aa-ad61-4e47-a533-6be5f25f7a37", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 920.513030] env[63088]: DEBUG oslo_vmware.api [None req-5a7bffe6-5d81-4eff-949f-17e8fd61a581 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Waiting for the task: (returnval){ [ 920.513030] env[63088]: value = "task-1285189" [ 920.513030] env[63088]: _type = "Task" [ 920.513030] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.514254] env[63088]: INFO nova.virt.block_device [None req-694d2129-1c60-4a36-8f0d-cb4e69a818c3 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Attempting to driver detach volume deedf893-eb6c-4ff0-ab0b-3e34abe88c8f from mountpoint /dev/sdb [ 920.514494] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-694d2129-1c60-4a36-8f0d-cb4e69a818c3 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Volume detach. Driver type: vmdk {{(pid=63088) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 920.514754] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-694d2129-1c60-4a36-8f0d-cb4e69a818c3 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-275902', 'volume_id': 'deedf893-eb6c-4ff0-ab0b-3e34abe88c8f', 'name': 'volume-deedf893-eb6c-4ff0-ab0b-3e34abe88c8f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69', 'attached_at': '', 'detached_at': '', 'volume_id': 'deedf893-eb6c-4ff0-ab0b-3e34abe88c8f', 'serial': 'deedf893-eb6c-4ff0-ab0b-3e34abe88c8f'} {{(pid=63088) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 920.515396] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f2:bc:c0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ae18b41f-e73c-44f1-83dd-467c080944f4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7e7a24aa-ad61-4e47-a533-6be5f25f7a37', 'vif_model': 'vmxnet3'}] {{(pid=63088) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 920.525137] env[63088]: DEBUG oslo.service.loopingcall [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 920.526105] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02c327cf-cd81-4b17-9c70-01c5d70687c0 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.529464] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3] Creating VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 920.533921] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-826f6083-48ef-4704-8e08-e6b0442dbac2 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.579213] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94fdc9f5-c85a-494c-a8d2-243ad6bcdc69 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.579271] env[63088]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 920.579271] env[63088]: value = "task-1285190" [ 920.579271] env[63088]: _type = "Task" [ 920.579271] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.579557] env[63088]: DEBUG oslo_vmware.api [None req-5a7bffe6-5d81-4eff-949f-17e8fd61a581 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285189, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.586673] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f863b1fa-dea6-4d4a-96ce-8116b4f878b4 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.593078] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285190, 'name': CreateVM_Task} progress is 15%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.615832] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c317e9f9-2607-4501-8077-c8951e5ae02b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.634209] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-694d2129-1c60-4a36-8f0d-cb4e69a818c3 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] The volume has not been displaced from its original location: [datastore1] volume-deedf893-eb6c-4ff0-ab0b-3e34abe88c8f/volume-deedf893-eb6c-4ff0-ab0b-3e34abe88c8f.vmdk. No consolidation needed. {{(pid=63088) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 920.639776] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-694d2129-1c60-4a36-8f0d-cb4e69a818c3 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Reconfiguring VM instance instance-00000043 to detach disk 2001 {{(pid=63088) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 920.640221] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9adec6cf-1c41-48b4-a53c-72560375d444 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.663367] env[63088]: DEBUG oslo_vmware.api [None req-694d2129-1c60-4a36-8f0d-cb4e69a818c3 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Waiting for the task: (returnval){ [ 920.663367] env[63088]: value = "task-1285191" [ 920.663367] env[63088]: _type = "Task" [ 920.663367] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.670828] env[63088]: DEBUG oslo_vmware.api [None req-694d2129-1c60-4a36-8f0d-cb4e69a818c3 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Task: {'id': task-1285191, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.776324] env[63088]: DEBUG nova.compute.manager [req-171cf0ae-c37b-461a-a478-cb1bc65167c8 req-d781bd23-1ab2-4d66-9bd9-47a69c126881 service nova] [instance: fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3] Received event network-changed-7e7a24aa-ad61-4e47-a533-6be5f25f7a37 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 920.776463] env[63088]: DEBUG nova.compute.manager [req-171cf0ae-c37b-461a-a478-cb1bc65167c8 req-d781bd23-1ab2-4d66-9bd9-47a69c126881 service nova] [instance: fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3] Refreshing instance network info cache due to event network-changed-7e7a24aa-ad61-4e47-a533-6be5f25f7a37. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 920.776697] env[63088]: DEBUG oslo_concurrency.lockutils [req-171cf0ae-c37b-461a-a478-cb1bc65167c8 req-d781bd23-1ab2-4d66-9bd9-47a69c126881 service nova] Acquiring lock "refresh_cache-fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 920.776844] env[63088]: DEBUG oslo_concurrency.lockutils [req-171cf0ae-c37b-461a-a478-cb1bc65167c8 req-d781bd23-1ab2-4d66-9bd9-47a69c126881 service nova] Acquired lock "refresh_cache-fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 920.777018] env[63088]: DEBUG nova.network.neutron [req-171cf0ae-c37b-461a-a478-cb1bc65167c8 req-d781bd23-1ab2-4d66-9bd9-47a69c126881 service nova] [instance: fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3] Refreshing network info cache for port 7e7a24aa-ad61-4e47-a533-6be5f25f7a37 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 920.831107] env[63088]: DEBUG oslo_concurrency.lockutils [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.274s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.831798] env[63088]: DEBUG nova.compute.manager [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 920.834923] env[63088]: DEBUG oslo_concurrency.lockutils [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.736s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.837892] env[63088]: INFO nova.compute.claims [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 921.018956] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquiring lock "2ad33254-8030-4454-b023-3359071f85de" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 921.019255] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Lock "2ad33254-8030-4454-b023-3359071f85de" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.036191] env[63088]: DEBUG oslo_vmware.api [None req-5a7bffe6-5d81-4eff-949f-17e8fd61a581 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285189, 'name': PowerOffVM_Task, 'duration_secs': 0.25959} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.036458] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a7bffe6-5d81-4eff-949f-17e8fd61a581 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 921.036632] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-5a7bffe6-5d81-4eff-949f-17e8fd61a581 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Unregistering the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 921.036890] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e8553f2c-7d58-4ced-ae65-c59c17a28431 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.090108] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285190, 'name': CreateVM_Task, 'duration_secs': 0.366151} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.090296] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3] Created VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 921.091138] env[63088]: DEBUG oslo_concurrency.lockutils [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 921.091138] env[63088]: DEBUG oslo_concurrency.lockutils [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 921.091453] env[63088]: DEBUG oslo_concurrency.lockutils [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 921.091717] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3b953e8a-7e8a-46f3-9bdf-12eeeedb54f2 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.096364] env[63088]: DEBUG oslo_vmware.api [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Waiting for the task: (returnval){ [ 921.096364] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52326dcd-a0f1-d7ed-5f8d-4c95cf2b8c98" [ 921.096364] env[63088]: _type = "Task" [ 921.096364] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.104412] env[63088]: DEBUG oslo_vmware.api [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52326dcd-a0f1-d7ed-5f8d-4c95cf2b8c98, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.171913] env[63088]: DEBUG oslo_vmware.api [None req-694d2129-1c60-4a36-8f0d-cb4e69a818c3 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Task: {'id': task-1285191, 'name': ReconfigVM_Task, 'duration_secs': 0.310989} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.172226] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-694d2129-1c60-4a36-8f0d-cb4e69a818c3 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Reconfigured VM instance instance-00000043 to detach disk 2001 {{(pid=63088) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 921.176946] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9dcc604d-fbe7-4bbf-b520-0693dec6a865 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.187393] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3d5df639-f016-4d69-8403-3ba3b2d15e1d tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Acquiring lock "e6b0ce08-d67f-458f-92f5-1e904d03a5ef" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 921.187635] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3d5df639-f016-4d69-8403-3ba3b2d15e1d tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Lock "e6b0ce08-d67f-458f-92f5-1e904d03a5ef" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.187846] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3d5df639-f016-4d69-8403-3ba3b2d15e1d tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Acquiring lock "e6b0ce08-d67f-458f-92f5-1e904d03a5ef-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 921.188081] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3d5df639-f016-4d69-8403-3ba3b2d15e1d tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Lock "e6b0ce08-d67f-458f-92f5-1e904d03a5ef-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.188272] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3d5df639-f016-4d69-8403-3ba3b2d15e1d tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Lock "e6b0ce08-d67f-458f-92f5-1e904d03a5ef-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 921.190645] env[63088]: INFO nova.compute.manager [None req-3d5df639-f016-4d69-8403-3ba3b2d15e1d tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Terminating instance [ 921.192717] env[63088]: DEBUG nova.compute.manager [None req-3d5df639-f016-4d69-8403-3ba3b2d15e1d tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 921.192916] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-3d5df639-f016-4d69-8403-3ba3b2d15e1d tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 921.193845] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1a1fc68-50c7-4931-b6dc-c9db2d1a98a4 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.198281] env[63088]: DEBUG oslo_vmware.api [None req-694d2129-1c60-4a36-8f0d-cb4e69a818c3 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Waiting for the task: (returnval){ [ 921.198281] env[63088]: value = "task-1285193" [ 921.198281] env[63088]: _type = "Task" [ 921.198281] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.207723] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-5a7bffe6-5d81-4eff-949f-17e8fd61a581 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Unregistered the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 921.207937] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-5a7bffe6-5d81-4eff-949f-17e8fd61a581 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Deleting contents of the VM from datastore datastore2 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 921.208142] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a7bffe6-5d81-4eff-949f-17e8fd61a581 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Deleting the datastore file [datastore2] 2269e520-08b6-433f-8fe9-7b84d91e02d7 {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 921.208422] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d5df639-f016-4d69-8403-3ba3b2d15e1d tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 921.209033] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e837bb0d-7bb1-4b84-ab63-d113243c2405 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.210920] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0996a3f8-993f-4ce7-99b4-fc2da12eacb9 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.215734] env[63088]: DEBUG oslo_vmware.api [None req-694d2129-1c60-4a36-8f0d-cb4e69a818c3 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Task: {'id': task-1285193, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.220573] env[63088]: DEBUG oslo_vmware.api [None req-5a7bffe6-5d81-4eff-949f-17e8fd61a581 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Waiting for the task: (returnval){ [ 921.220573] env[63088]: value = "task-1285194" [ 921.220573] env[63088]: _type = "Task" [ 921.220573] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.221963] env[63088]: DEBUG oslo_vmware.api [None req-3d5df639-f016-4d69-8403-3ba3b2d15e1d tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Waiting for the task: (returnval){ [ 921.221963] env[63088]: value = "task-1285195" [ 921.221963] env[63088]: _type = "Task" [ 921.221963] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.232888] env[63088]: DEBUG oslo_vmware.api [None req-3d5df639-f016-4d69-8403-3ba3b2d15e1d tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Task: {'id': task-1285195, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.235986] env[63088]: DEBUG oslo_vmware.api [None req-5a7bffe6-5d81-4eff-949f-17e8fd61a581 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285194, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.345036] env[63088]: DEBUG nova.compute.utils [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 921.346702] env[63088]: DEBUG nova.compute.manager [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 921.346865] env[63088]: DEBUG nova.network.neutron [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 921.411662] env[63088]: DEBUG nova.policy [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '139b65d9cf314b3db1d685d63bc19bba', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e48f62dee9ad4e9b94b67bd871db5c63', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 921.521864] env[63088]: DEBUG nova.compute.manager [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2ad33254-8030-4454-b023-3359071f85de] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 921.537140] env[63088]: DEBUG nova.network.neutron [req-171cf0ae-c37b-461a-a478-cb1bc65167c8 req-d781bd23-1ab2-4d66-9bd9-47a69c126881 service nova] [instance: fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3] Updated VIF entry in instance network info cache for port 7e7a24aa-ad61-4e47-a533-6be5f25f7a37. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 921.537660] env[63088]: DEBUG nova.network.neutron [req-171cf0ae-c37b-461a-a478-cb1bc65167c8 req-d781bd23-1ab2-4d66-9bd9-47a69c126881 service nova] [instance: fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3] Updating instance_info_cache with network_info: [{"id": "7e7a24aa-ad61-4e47-a533-6be5f25f7a37", "address": "fa:16:3e:f2:bc:c0", "network": {"id": "307b966a-d9e1-40fd-9313-1ad94c734308", "bridge": "br-int", "label": "tempest-ServersTestJSON-1907249197-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "03e20d606e654362acbe2b36fe499ae3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae18b41f-e73c-44f1-83dd-467c080944f4", "external-id": "nsx-vlan-transportzone-653", "segmentation_id": 653, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e7a24aa-ad", "ovs_interfaceid": "7e7a24aa-ad61-4e47-a533-6be5f25f7a37", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 921.611084] env[63088]: DEBUG oslo_vmware.api [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52326dcd-a0f1-d7ed-5f8d-4c95cf2b8c98, 'name': SearchDatastore_Task, 'duration_secs': 0.011358} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.611084] env[63088]: DEBUG oslo_concurrency.lockutils [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 921.611084] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3] Processing image 1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 921.611084] env[63088]: DEBUG oslo_concurrency.lockutils [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 921.611084] env[63088]: DEBUG oslo_concurrency.lockutils [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 921.611084] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 921.611084] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-373f2de5-3bda-4761-9db6-ca6496dd0ee6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.624629] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 921.624629] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 921.624998] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7d4fb966-c769-45e8-9baf-cb692f70f998 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.631742] env[63088]: DEBUG oslo_vmware.api [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Waiting for the task: (returnval){ [ 921.631742] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]5279fcad-933c-45f4-bfe3-73d2a2a5a999" [ 921.631742] env[63088]: _type = "Task" [ 921.631742] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.640371] env[63088]: DEBUG oslo_vmware.api [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]5279fcad-933c-45f4-bfe3-73d2a2a5a999, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.696358] env[63088]: DEBUG nova.network.neutron [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] Successfully created port: 88203cff-5be6-4261-a168-197e6be13cc3 {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 921.708503] env[63088]: DEBUG oslo_vmware.api [None req-694d2129-1c60-4a36-8f0d-cb4e69a818c3 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Task: {'id': task-1285193, 'name': ReconfigVM_Task, 'duration_secs': 0.154506} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.708826] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-694d2129-1c60-4a36-8f0d-cb4e69a818c3 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-275902', 'volume_id': 'deedf893-eb6c-4ff0-ab0b-3e34abe88c8f', 'name': 'volume-deedf893-eb6c-4ff0-ab0b-3e34abe88c8f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69', 'attached_at': '', 'detached_at': '', 'volume_id': 'deedf893-eb6c-4ff0-ab0b-3e34abe88c8f', 'serial': 'deedf893-eb6c-4ff0-ab0b-3e34abe88c8f'} {{(pid=63088) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 921.734857] env[63088]: DEBUG oslo_vmware.api [None req-5a7bffe6-5d81-4eff-949f-17e8fd61a581 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285194, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.22654} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.737971] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a7bffe6-5d81-4eff-949f-17e8fd61a581 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 921.738217] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-5a7bffe6-5d81-4eff-949f-17e8fd61a581 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Deleted contents of the VM from datastore datastore2 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 921.738410] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-5a7bffe6-5d81-4eff-949f-17e8fd61a581 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 921.738589] env[63088]: INFO nova.compute.manager [None req-5a7bffe6-5d81-4eff-949f-17e8fd61a581 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Took 1.25 seconds to destroy the instance on the hypervisor. [ 921.738971] env[63088]: DEBUG oslo.service.loopingcall [None req-5a7bffe6-5d81-4eff-949f-17e8fd61a581 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 921.739484] env[63088]: DEBUG oslo_vmware.api [None req-3d5df639-f016-4d69-8403-3ba3b2d15e1d tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Task: {'id': task-1285195, 'name': PowerOffVM_Task, 'duration_secs': 0.212485} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.739720] env[63088]: DEBUG nova.compute.manager [-] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 921.739815] env[63088]: DEBUG nova.network.neutron [-] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 921.741571] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d5df639-f016-4d69-8403-3ba3b2d15e1d tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 921.741752] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-3d5df639-f016-4d69-8403-3ba3b2d15e1d tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Unregistering the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 921.742018] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d1c123bc-da9d-4846-805c-b4c36f69a51b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.852147] env[63088]: DEBUG nova.compute.manager [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 921.854692] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-3d5df639-f016-4d69-8403-3ba3b2d15e1d tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Unregistered the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 921.858027] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-3d5df639-f016-4d69-8403-3ba3b2d15e1d tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Deleting contents of the VM from datastore datastore2 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 921.858027] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d5df639-f016-4d69-8403-3ba3b2d15e1d tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Deleting the datastore file [datastore2] e6b0ce08-d67f-458f-92f5-1e904d03a5ef {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 921.858027] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bf0b5c97-5882-4002-b976-cfee17310fae {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.862362] env[63088]: DEBUG oslo_vmware.api [None req-3d5df639-f016-4d69-8403-3ba3b2d15e1d tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Waiting for the task: (returnval){ [ 921.862362] env[63088]: value = "task-1285197" [ 921.862362] env[63088]: _type = "Task" [ 921.862362] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.871459] env[63088]: DEBUG oslo_vmware.api [None req-3d5df639-f016-4d69-8403-3ba3b2d15e1d tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Task: {'id': task-1285197, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.040307] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 922.040794] env[63088]: DEBUG oslo_concurrency.lockutils [req-171cf0ae-c37b-461a-a478-cb1bc65167c8 req-d781bd23-1ab2-4d66-9bd9-47a69c126881 service nova] Releasing lock "refresh_cache-fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 922.076021] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13087fb1-c841-4a50-94ae-5241c0226f42 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.081833] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-000ab109-4770-4157-9448-6abf4c33cf5f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.112936] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab2bb615-1d98-4045-bbcb-be6b075b40e8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.119977] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b42e4848-5a4c-4a98-bf83-9ea652aea09d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.132833] env[63088]: DEBUG nova.compute.provider_tree [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 922.141362] env[63088]: DEBUG oslo_vmware.api [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]5279fcad-933c-45f4-bfe3-73d2a2a5a999, 'name': SearchDatastore_Task, 'duration_secs': 0.009578} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.142680] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-040e33aa-23a4-4b74-8919-aeade915c807 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.148547] env[63088]: DEBUG oslo_vmware.api [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Waiting for the task: (returnval){ [ 922.148547] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]521dbe1c-eed2-cd1c-0408-d22c39dbd35a" [ 922.148547] env[63088]: _type = "Task" [ 922.148547] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.155689] env[63088]: DEBUG oslo_vmware.api [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]521dbe1c-eed2-cd1c-0408-d22c39dbd35a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.256359] env[63088]: DEBUG nova.objects.instance [None req-694d2129-1c60-4a36-8f0d-cb4e69a818c3 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Lazy-loading 'flavor' on Instance uuid dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 922.372107] env[63088]: DEBUG oslo_vmware.api [None req-3d5df639-f016-4d69-8403-3ba3b2d15e1d tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Task: {'id': task-1285197, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.160574} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.372372] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d5df639-f016-4d69-8403-3ba3b2d15e1d tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 922.372559] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-3d5df639-f016-4d69-8403-3ba3b2d15e1d tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Deleted contents of the VM from datastore datastore2 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 922.372737] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-3d5df639-f016-4d69-8403-3ba3b2d15e1d tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 922.372911] env[63088]: INFO nova.compute.manager [None req-3d5df639-f016-4d69-8403-3ba3b2d15e1d tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Took 1.18 seconds to destroy the instance on the hypervisor. [ 922.373198] env[63088]: DEBUG oslo.service.loopingcall [None req-3d5df639-f016-4d69-8403-3ba3b2d15e1d tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 922.373409] env[63088]: DEBUG nova.compute.manager [-] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 922.373505] env[63088]: DEBUG nova.network.neutron [-] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 922.523234] env[63088]: DEBUG nova.network.neutron [-] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 922.637827] env[63088]: DEBUG nova.scheduler.client.report [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 922.666080] env[63088]: DEBUG oslo_vmware.api [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]521dbe1c-eed2-cd1c-0408-d22c39dbd35a, 'name': SearchDatastore_Task, 'duration_secs': 0.009189} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.666080] env[63088]: DEBUG oslo_concurrency.lockutils [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 922.666080] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3/fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 922.666080] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-40032adc-5b81-487b-aa6f-6ffb1835c70f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.676239] env[63088]: DEBUG oslo_vmware.api [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Waiting for the task: (returnval){ [ 922.676239] env[63088]: value = "task-1285198" [ 922.676239] env[63088]: _type = "Task" [ 922.676239] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.687521] env[63088]: DEBUG oslo_vmware.api [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285198, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.804803] env[63088]: DEBUG nova.compute.manager [req-b8a6ff7c-59ac-44a8-a101-d98d1def6453 req-dcdfe3df-d6a9-4c87-bc9d-84befd57d227 service nova] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Received event network-vif-deleted-1fb0e170-03ab-4ee4-9f1f-c5c6008c823d {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 922.805048] env[63088]: DEBUG nova.compute.manager [req-b8a6ff7c-59ac-44a8-a101-d98d1def6453 req-dcdfe3df-d6a9-4c87-bc9d-84befd57d227 service nova] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Received event network-vif-deleted-63d8dbbe-4550-4d8d-8e85-9a9eadc81d82 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 922.805250] env[63088]: INFO nova.compute.manager [req-b8a6ff7c-59ac-44a8-a101-d98d1def6453 req-dcdfe3df-d6a9-4c87-bc9d-84befd57d227 service nova] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Neutron deleted interface 63d8dbbe-4550-4d8d-8e85-9a9eadc81d82; detaching it from the instance and deleting it from the info cache [ 922.805435] env[63088]: DEBUG nova.network.neutron [req-b8a6ff7c-59ac-44a8-a101-d98d1def6453 req-dcdfe3df-d6a9-4c87-bc9d-84befd57d227 service nova] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 922.861658] env[63088]: DEBUG nova.compute.manager [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 922.889370] env[63088]: DEBUG nova.virt.hardware [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 922.889675] env[63088]: DEBUG nova.virt.hardware [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 922.889893] env[63088]: DEBUG nova.virt.hardware [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 922.890287] env[63088]: DEBUG nova.virt.hardware [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 922.890505] env[63088]: DEBUG nova.virt.hardware [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 922.890882] env[63088]: DEBUG nova.virt.hardware [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 922.891346] env[63088]: DEBUG nova.virt.hardware [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 922.891573] env[63088]: DEBUG nova.virt.hardware [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 922.893366] env[63088]: DEBUG nova.virt.hardware [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 922.893366] env[63088]: DEBUG nova.virt.hardware [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 922.893366] env[63088]: DEBUG nova.virt.hardware [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 922.893366] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d38e57a-f35b-4eef-af3c-5ea18c1b3e23 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.903429] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58570cc2-8b6b-42a2-b611-31fc3111d706 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.026757] env[63088]: INFO nova.compute.manager [-] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Took 1.29 seconds to deallocate network for instance. [ 923.137824] env[63088]: DEBUG nova.compute.manager [req-4c31b0a8-3e87-4e87-80f8-755b2379e3bd req-477748be-42c9-4135-93ed-983ef8dfee83 service nova] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] Received event network-vif-plugged-88203cff-5be6-4261-a168-197e6be13cc3 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 923.138144] env[63088]: DEBUG oslo_concurrency.lockutils [req-4c31b0a8-3e87-4e87-80f8-755b2379e3bd req-477748be-42c9-4135-93ed-983ef8dfee83 service nova] Acquiring lock "72c07d1e-cbb6-4875-9b7f-2649aa25eee5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.138391] env[63088]: DEBUG oslo_concurrency.lockutils [req-4c31b0a8-3e87-4e87-80f8-755b2379e3bd req-477748be-42c9-4135-93ed-983ef8dfee83 service nova] Lock "72c07d1e-cbb6-4875-9b7f-2649aa25eee5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.138620] env[63088]: DEBUG oslo_concurrency.lockutils [req-4c31b0a8-3e87-4e87-80f8-755b2379e3bd req-477748be-42c9-4135-93ed-983ef8dfee83 service nova] Lock "72c07d1e-cbb6-4875-9b7f-2649aa25eee5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.138813] env[63088]: DEBUG nova.compute.manager [req-4c31b0a8-3e87-4e87-80f8-755b2379e3bd req-477748be-42c9-4135-93ed-983ef8dfee83 service nova] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] No waiting events found dispatching network-vif-plugged-88203cff-5be6-4261-a168-197e6be13cc3 {{(pid=63088) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 923.138989] env[63088]: WARNING nova.compute.manager [req-4c31b0a8-3e87-4e87-80f8-755b2379e3bd req-477748be-42c9-4135-93ed-983ef8dfee83 service nova] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] Received unexpected event network-vif-plugged-88203cff-5be6-4261-a168-197e6be13cc3 for instance with vm_state building and task_state spawning. [ 923.146506] env[63088]: DEBUG oslo_concurrency.lockutils [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.312s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.147065] env[63088]: DEBUG nova.compute.manager [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 923.150387] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6545a09f-2d02-4a01-bf68-ba1a748f4bee tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.030s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.150387] env[63088]: DEBUG nova.objects.instance [None req-6545a09f-2d02-4a01-bf68-ba1a748f4bee tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Lazy-loading 'resources' on Instance uuid b50db882-598c-488f-b935-34f55f655642 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 923.174481] env[63088]: DEBUG nova.network.neutron [-] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 923.187479] env[63088]: DEBUG oslo_vmware.api [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285198, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.485594} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.188808] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3/fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 923.189051] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3] Extending root virtual disk to 1048576 {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 923.189332] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-643ede20-b2f5-44d6-b89c-34f21d3a513b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.198641] env[63088]: DEBUG oslo_vmware.api [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Waiting for the task: (returnval){ [ 923.198641] env[63088]: value = "task-1285199" [ 923.198641] env[63088]: _type = "Task" [ 923.198641] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.207322] env[63088]: DEBUG oslo_vmware.api [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285199, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.245693] env[63088]: DEBUG nova.network.neutron [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] Successfully updated port: 88203cff-5be6-4261-a168-197e6be13cc3 {{(pid=63088) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 923.260972] env[63088]: DEBUG oslo_concurrency.lockutils [None req-8f0c8457-3b0e-4eee-96ac-fd47d1f0935f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Acquiring lock "dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.262817] env[63088]: DEBUG oslo_concurrency.lockutils [None req-694d2129-1c60-4a36-8f0d-cb4e69a818c3 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Lock "dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.301s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.264165] env[63088]: DEBUG oslo_concurrency.lockutils [None req-8f0c8457-3b0e-4eee-96ac-fd47d1f0935f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Lock "dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.003s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.307933] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a81b1eb8-b796-4fed-ba47-36e2099d012c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.319272] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b83d8eb6-3bec-4d98-b73b-884f0388b87b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.343286] env[63088]: DEBUG nova.compute.manager [req-b8a6ff7c-59ac-44a8-a101-d98d1def6453 req-dcdfe3df-d6a9-4c87-bc9d-84befd57d227 service nova] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Detach interface failed, port_id=63d8dbbe-4550-4d8d-8e85-9a9eadc81d82, reason: Instance e6b0ce08-d67f-458f-92f5-1e904d03a5ef could not be found. {{(pid=63088) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 923.534306] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5a7bffe6-5d81-4eff-949f-17e8fd61a581 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.653436] env[63088]: DEBUG nova.compute.utils [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 923.654900] env[63088]: DEBUG nova.compute.manager [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 923.655080] env[63088]: DEBUG nova.network.neutron [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 923.677547] env[63088]: INFO nova.compute.manager [-] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Took 1.30 seconds to deallocate network for instance. [ 923.698141] env[63088]: DEBUG nova.policy [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '93f1d334e71a4a99ba300c3ee6d70243', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'aa3d24a1a6c0430985fd80365d986ee1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 923.710192] env[63088]: DEBUG oslo_vmware.api [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285199, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061162} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.710482] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3] Extended root virtual disk {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 923.711432] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d3d1559-5601-46c8-835d-9febe0beb578 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.735171] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3] Reconfiguring VM instance instance-00000050 to attach disk [datastore1] fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3/fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 923.739439] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bde1d165-e565-41ee-8f2d-aaa83fe0d14a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.753815] env[63088]: DEBUG oslo_concurrency.lockutils [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquiring lock "refresh_cache-72c07d1e-cbb6-4875-9b7f-2649aa25eee5" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 923.753957] env[63088]: DEBUG oslo_concurrency.lockutils [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquired lock "refresh_cache-72c07d1e-cbb6-4875-9b7f-2649aa25eee5" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.754134] env[63088]: DEBUG nova.network.neutron [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 923.760822] env[63088]: DEBUG oslo_vmware.api [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Waiting for the task: (returnval){ [ 923.760822] env[63088]: value = "task-1285200" [ 923.760822] env[63088]: _type = "Task" [ 923.760822] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.767327] env[63088]: INFO nova.compute.manager [None req-8f0c8457-3b0e-4eee-96ac-fd47d1f0935f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Detaching volume 51f9c2dc-ac18-40cf-87db-5359fcb203a8 [ 923.774470] env[63088]: DEBUG oslo_vmware.api [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285200, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.810159] env[63088]: INFO nova.virt.block_device [None req-8f0c8457-3b0e-4eee-96ac-fd47d1f0935f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Attempting to driver detach volume 51f9c2dc-ac18-40cf-87db-5359fcb203a8 from mountpoint /dev/sdc [ 923.810432] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-8f0c8457-3b0e-4eee-96ac-fd47d1f0935f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Volume detach. Driver type: vmdk {{(pid=63088) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 923.810841] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-8f0c8457-3b0e-4eee-96ac-fd47d1f0935f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-275907', 'volume_id': '51f9c2dc-ac18-40cf-87db-5359fcb203a8', 'name': 'volume-51f9c2dc-ac18-40cf-87db-5359fcb203a8', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69', 'attached_at': '', 'detached_at': '', 'volume_id': '51f9c2dc-ac18-40cf-87db-5359fcb203a8', 'serial': '51f9c2dc-ac18-40cf-87db-5359fcb203a8'} {{(pid=63088) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 923.811716] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7d0d127-1f63-499b-8a30-1f895236144b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.837670] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dda42439-5b06-4dba-a460-27cd4908e122 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.844883] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0bd57c3-e493-4ae0-ac63-61d08810eac4 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.868631] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a07a24c-511d-40f5-b3d3-7cd835bf6519 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.883965] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-8f0c8457-3b0e-4eee-96ac-fd47d1f0935f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] The volume has not been displaced from its original location: [datastore1] volume-51f9c2dc-ac18-40cf-87db-5359fcb203a8/volume-51f9c2dc-ac18-40cf-87db-5359fcb203a8.vmdk. No consolidation needed. {{(pid=63088) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 923.890925] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-8f0c8457-3b0e-4eee-96ac-fd47d1f0935f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Reconfiguring VM instance instance-00000043 to detach disk 2002 {{(pid=63088) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 923.893114] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5af57af7-c80a-437d-a356-1d5b6634ab02 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.912812] env[63088]: DEBUG oslo_vmware.api [None req-8f0c8457-3b0e-4eee-96ac-fd47d1f0935f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Waiting for the task: (returnval){ [ 923.912812] env[63088]: value = "task-1285201" [ 923.912812] env[63088]: _type = "Task" [ 923.912812] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.924354] env[63088]: DEBUG oslo_vmware.api [None req-8f0c8457-3b0e-4eee-96ac-fd47d1f0935f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Task: {'id': task-1285201, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.958162] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75cb1739-0330-4c31-9f4e-8deceae7b353 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.965605] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-042cbee0-9c59-4c45-8176-a2a9656ce38f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.996111] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10138d41-7b63-468b-b0d2-1c85a12af28d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.003624] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1e63ddf-21e1-413b-809f-4aa92b5f36f4 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.017013] env[63088]: DEBUG nova.compute.provider_tree [None req-6545a09f-2d02-4a01-bf68-ba1a748f4bee tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 924.019243] env[63088]: DEBUG nova.network.neutron [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Successfully created port: 81fc2dda-617e-4e4c-aa91-7e7020003995 {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 924.161015] env[63088]: DEBUG nova.compute.manager [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 924.186327] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3d5df639-f016-4d69-8403-3ba3b2d15e1d tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 924.278594] env[63088]: DEBUG oslo_vmware.api [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285200, 'name': ReconfigVM_Task, 'duration_secs': 0.325678} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.278594] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3] Reconfigured VM instance instance-00000050 to attach disk [datastore1] fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3/fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 924.278784] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-45305000-caf8-4cc0-9772-65f4ff5dc09c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.284994] env[63088]: DEBUG oslo_vmware.api [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Waiting for the task: (returnval){ [ 924.284994] env[63088]: value = "task-1285202" [ 924.284994] env[63088]: _type = "Task" [ 924.284994] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.293749] env[63088]: DEBUG oslo_vmware.api [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285202, 'name': Rename_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.295893] env[63088]: DEBUG nova.network.neutron [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 924.422349] env[63088]: DEBUG oslo_vmware.api [None req-8f0c8457-3b0e-4eee-96ac-fd47d1f0935f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Task: {'id': task-1285201, 'name': ReconfigVM_Task, 'duration_secs': 0.235134} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.422566] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-8f0c8457-3b0e-4eee-96ac-fd47d1f0935f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Reconfigured VM instance instance-00000043 to detach disk 2002 {{(pid=63088) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 924.427171] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-73da31d0-7abc-4c61-9346-c69fedfbb6b3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.438190] env[63088]: DEBUG nova.network.neutron [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] Updating instance_info_cache with network_info: [{"id": "88203cff-5be6-4261-a168-197e6be13cc3", "address": "fa:16:3e:e8:83:1f", "network": {"id": "5b801381-6bb3-45cd-b81f-92b5d757e080", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1481457494-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e48f62dee9ad4e9b94b67bd871db5c63", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "051f343d-ac4f-4070-a26d-467603122c81", "external-id": "nsx-vlan-transportzone-277", "segmentation_id": 277, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap88203cff-5b", "ovs_interfaceid": "88203cff-5be6-4261-a168-197e6be13cc3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 924.444150] env[63088]: DEBUG oslo_vmware.api [None req-8f0c8457-3b0e-4eee-96ac-fd47d1f0935f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Waiting for the task: (returnval){ [ 924.444150] env[63088]: value = "task-1285203" [ 924.444150] env[63088]: _type = "Task" [ 924.444150] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.452444] env[63088]: DEBUG oslo_vmware.api [None req-8f0c8457-3b0e-4eee-96ac-fd47d1f0935f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Task: {'id': task-1285203, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.522719] env[63088]: DEBUG nova.scheduler.client.report [None req-6545a09f-2d02-4a01-bf68-ba1a748f4bee tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 924.795534] env[63088]: DEBUG oslo_vmware.api [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285202, 'name': Rename_Task, 'duration_secs': 0.141804} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.795863] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 924.796165] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-10b0a1b7-f098-4195-a878-c0853f2dc5f3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.803105] env[63088]: DEBUG oslo_vmware.api [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Waiting for the task: (returnval){ [ 924.803105] env[63088]: value = "task-1285204" [ 924.803105] env[63088]: _type = "Task" [ 924.803105] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.816904] env[63088]: DEBUG oslo_vmware.api [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285204, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.941437] env[63088]: DEBUG oslo_concurrency.lockutils [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Releasing lock "refresh_cache-72c07d1e-cbb6-4875-9b7f-2649aa25eee5" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 924.941799] env[63088]: DEBUG nova.compute.manager [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] Instance network_info: |[{"id": "88203cff-5be6-4261-a168-197e6be13cc3", "address": "fa:16:3e:e8:83:1f", "network": {"id": "5b801381-6bb3-45cd-b81f-92b5d757e080", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1481457494-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e48f62dee9ad4e9b94b67bd871db5c63", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "051f343d-ac4f-4070-a26d-467603122c81", "external-id": "nsx-vlan-transportzone-277", "segmentation_id": 277, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap88203cff-5b", "ovs_interfaceid": "88203cff-5be6-4261-a168-197e6be13cc3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 924.942367] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e8:83:1f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '051f343d-ac4f-4070-a26d-467603122c81', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '88203cff-5be6-4261-a168-197e6be13cc3', 'vif_model': 'vmxnet3'}] {{(pid=63088) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 924.951269] env[63088]: DEBUG oslo.service.loopingcall [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 924.951608] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] Creating VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 924.954906] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c1653412-2c95-433b-9777-f332c5dca51d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.974368] env[63088]: DEBUG oslo_vmware.api [None req-8f0c8457-3b0e-4eee-96ac-fd47d1f0935f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Task: {'id': task-1285203, 'name': ReconfigVM_Task, 'duration_secs': 0.119785} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.975602] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-8f0c8457-3b0e-4eee-96ac-fd47d1f0935f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-275907', 'volume_id': '51f9c2dc-ac18-40cf-87db-5359fcb203a8', 'name': 'volume-51f9c2dc-ac18-40cf-87db-5359fcb203a8', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69', 'attached_at': '', 'detached_at': '', 'volume_id': '51f9c2dc-ac18-40cf-87db-5359fcb203a8', 'serial': '51f9c2dc-ac18-40cf-87db-5359fcb203a8'} {{(pid=63088) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 924.977689] env[63088]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 924.977689] env[63088]: value = "task-1285205" [ 924.977689] env[63088]: _type = "Task" [ 924.977689] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.985998] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285205, 'name': CreateVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.027976] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6545a09f-2d02-4a01-bf68-ba1a748f4bee tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.878s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.030546] env[63088]: DEBUG oslo_concurrency.lockutils [None req-cd65c449-e671-45aa-a201-5e68b37dd782 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.867s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.030789] env[63088]: DEBUG nova.objects.instance [None req-cd65c449-e671-45aa-a201-5e68b37dd782 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Lazy-loading 'resources' on Instance uuid 2964db35-7357-40a7-b4e6-7e2595549f5b {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 925.052692] env[63088]: INFO nova.scheduler.client.report [None req-6545a09f-2d02-4a01-bf68-ba1a748f4bee tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Deleted allocations for instance b50db882-598c-488f-b935-34f55f655642 [ 925.162817] env[63088]: DEBUG nova.compute.manager [req-7911e47f-7121-4007-a226-e79eba1e0722 req-93a43156-e4a6-4c11-a66c-b23c9c1ed78c service nova] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] Received event network-changed-88203cff-5be6-4261-a168-197e6be13cc3 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 925.163045] env[63088]: DEBUG nova.compute.manager [req-7911e47f-7121-4007-a226-e79eba1e0722 req-93a43156-e4a6-4c11-a66c-b23c9c1ed78c service nova] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] Refreshing instance network info cache due to event network-changed-88203cff-5be6-4261-a168-197e6be13cc3. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 925.163324] env[63088]: DEBUG oslo_concurrency.lockutils [req-7911e47f-7121-4007-a226-e79eba1e0722 req-93a43156-e4a6-4c11-a66c-b23c9c1ed78c service nova] Acquiring lock "refresh_cache-72c07d1e-cbb6-4875-9b7f-2649aa25eee5" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 925.163499] env[63088]: DEBUG oslo_concurrency.lockutils [req-7911e47f-7121-4007-a226-e79eba1e0722 req-93a43156-e4a6-4c11-a66c-b23c9c1ed78c service nova] Acquired lock "refresh_cache-72c07d1e-cbb6-4875-9b7f-2649aa25eee5" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.163751] env[63088]: DEBUG nova.network.neutron [req-7911e47f-7121-4007-a226-e79eba1e0722 req-93a43156-e4a6-4c11-a66c-b23c9c1ed78c service nova] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] Refreshing network info cache for port 88203cff-5be6-4261-a168-197e6be13cc3 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 925.174134] env[63088]: DEBUG nova.compute.manager [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 925.201086] env[63088]: DEBUG nova.virt.hardware [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 925.201468] env[63088]: DEBUG nova.virt.hardware [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 925.201656] env[63088]: DEBUG nova.virt.hardware [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 925.201918] env[63088]: DEBUG nova.virt.hardware [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 925.202143] env[63088]: DEBUG nova.virt.hardware [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 925.202322] env[63088]: DEBUG nova.virt.hardware [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 925.202546] env[63088]: DEBUG nova.virt.hardware [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 925.202714] env[63088]: DEBUG nova.virt.hardware [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 925.202893] env[63088]: DEBUG nova.virt.hardware [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 925.203179] env[63088]: DEBUG nova.virt.hardware [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 925.203496] env[63088]: DEBUG nova.virt.hardware [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 925.205222] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a372b0c5-003d-461a-830e-ed31ff2e4b43 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.214563] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a56849b-7dc8-4f1a-a8ea-208285024aad {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.312872] env[63088]: DEBUG oslo_vmware.api [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285204, 'name': PowerOnVM_Task, 'duration_secs': 0.440258} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.313472] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 925.313472] env[63088]: INFO nova.compute.manager [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3] Took 7.13 seconds to spawn the instance on the hypervisor. [ 925.313625] env[63088]: DEBUG nova.compute.manager [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 925.314391] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b802a1f1-ddbb-4c7c-a147-bf3d46a6a723 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.488174] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285205, 'name': CreateVM_Task, 'duration_secs': 0.343402} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.488344] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] Created VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 925.489016] env[63088]: DEBUG oslo_concurrency.lockutils [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 925.489205] env[63088]: DEBUG oslo_concurrency.lockutils [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.489530] env[63088]: DEBUG oslo_concurrency.lockutils [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 925.489794] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d126c3c1-7766-47ff-a913-edbf56935c48 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.494192] env[63088]: DEBUG oslo_vmware.api [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 925.494192] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52a80d91-ac80-63c3-e6a2-81a5faf23e7f" [ 925.494192] env[63088]: _type = "Task" [ 925.494192] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.501768] env[63088]: DEBUG oslo_vmware.api [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52a80d91-ac80-63c3-e6a2-81a5faf23e7f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.521096] env[63088]: DEBUG nova.network.neutron [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Successfully updated port: 81fc2dda-617e-4e4c-aa91-7e7020003995 {{(pid=63088) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 925.529644] env[63088]: DEBUG nova.objects.instance [None req-8f0c8457-3b0e-4eee-96ac-fd47d1f0935f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Lazy-loading 'flavor' on Instance uuid dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 925.560862] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6545a09f-2d02-4a01-bf68-ba1a748f4bee tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Lock "b50db882-598c-488f-b935-34f55f655642" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.395s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.697628] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6c49aa5-6b14-4122-ae40-ac01760498ec {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.704632] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9939d05-9eae-4a63-991d-e0d74669857a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.743613] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aec68fa6-9d36-47cb-88e9-bd0c4a1a3508 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.751497] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4413db5e-649b-47b9-8844-ba59f08cb26f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.766039] env[63088]: DEBUG nova.compute.provider_tree [None req-cd65c449-e671-45aa-a201-5e68b37dd782 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 925.831563] env[63088]: INFO nova.compute.manager [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3] Took 16.57 seconds to build instance. [ 925.947613] env[63088]: DEBUG nova.network.neutron [req-7911e47f-7121-4007-a226-e79eba1e0722 req-93a43156-e4a6-4c11-a66c-b23c9c1ed78c service nova] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] Updated VIF entry in instance network info cache for port 88203cff-5be6-4261-a168-197e6be13cc3. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 925.948126] env[63088]: DEBUG nova.network.neutron [req-7911e47f-7121-4007-a226-e79eba1e0722 req-93a43156-e4a6-4c11-a66c-b23c9c1ed78c service nova] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] Updating instance_info_cache with network_info: [{"id": "88203cff-5be6-4261-a168-197e6be13cc3", "address": "fa:16:3e:e8:83:1f", "network": {"id": "5b801381-6bb3-45cd-b81f-92b5d757e080", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1481457494-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e48f62dee9ad4e9b94b67bd871db5c63", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "051f343d-ac4f-4070-a26d-467603122c81", "external-id": "nsx-vlan-transportzone-277", "segmentation_id": 277, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap88203cff-5b", "ovs_interfaceid": "88203cff-5be6-4261-a168-197e6be13cc3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 926.006691] env[63088]: DEBUG oslo_vmware.api [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52a80d91-ac80-63c3-e6a2-81a5faf23e7f, 'name': SearchDatastore_Task, 'duration_secs': 0.010785} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.006691] env[63088]: DEBUG oslo_concurrency.lockutils [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 926.006691] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] Processing image 1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 926.006925] env[63088]: DEBUG oslo_concurrency.lockutils [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 926.006925] env[63088]: DEBUG oslo_concurrency.lockutils [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 926.007113] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 926.007376] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7adfdf08-f6ab-4421-867c-4d7c2d8abff3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.015355] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 926.015526] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 926.016230] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4c3284a9-6144-47cd-9174-e99e0df169b1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.020989] env[63088]: DEBUG oslo_vmware.api [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 926.020989] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52cd5508-3669-a6aa-c7f5-656917123fa2" [ 926.020989] env[63088]: _type = "Task" [ 926.020989] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.024423] env[63088]: DEBUG oslo_concurrency.lockutils [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquiring lock "refresh_cache-f582da15-dfc3-45e1-a995-9dd0c9533869" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 926.024423] env[63088]: DEBUG oslo_concurrency.lockutils [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquired lock "refresh_cache-f582da15-dfc3-45e1-a995-9dd0c9533869" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 926.024587] env[63088]: DEBUG nova.network.neutron [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 926.029831] env[63088]: DEBUG oslo_vmware.api [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52cd5508-3669-a6aa-c7f5-656917123fa2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.269099] env[63088]: DEBUG nova.scheduler.client.report [None req-cd65c449-e671-45aa-a201-5e68b37dd782 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 926.334221] env[63088]: DEBUG oslo_concurrency.lockutils [None req-633ef0be-389a-4b22-b1bb-3fd9ea851e97 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lock "fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.084s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.357743] env[63088]: DEBUG oslo_concurrency.lockutils [None req-dcb8cf57-4815-4af4-b251-04f53799c533 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquiring lock "7b6aadb7-e34b-42b7-b69f-370434f5b665" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.358069] env[63088]: DEBUG oslo_concurrency.lockutils [None req-dcb8cf57-4815-4af4-b251-04f53799c533 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lock "7b6aadb7-e34b-42b7-b69f-370434f5b665" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.358332] env[63088]: INFO nova.compute.manager [None req-dcb8cf57-4815-4af4-b251-04f53799c533 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Rebooting instance [ 926.451145] env[63088]: DEBUG oslo_concurrency.lockutils [req-7911e47f-7121-4007-a226-e79eba1e0722 req-93a43156-e4a6-4c11-a66c-b23c9c1ed78c service nova] Releasing lock "refresh_cache-72c07d1e-cbb6-4875-9b7f-2649aa25eee5" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 926.532396] env[63088]: DEBUG oslo_vmware.api [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52cd5508-3669-a6aa-c7f5-656917123fa2, 'name': SearchDatastore_Task, 'duration_secs': 0.008823} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.533227] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6c5dca10-f647-4a84-b870-c334c0030c86 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.538277] env[63088]: DEBUG oslo_vmware.api [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 926.538277] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]525d3e30-9b77-8b36-ef1f-bc2562d4bc2c" [ 926.538277] env[63088]: _type = "Task" [ 926.538277] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.538795] env[63088]: DEBUG oslo_concurrency.lockutils [None req-8f0c8457-3b0e-4eee-96ac-fd47d1f0935f tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Lock "dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.275s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.550379] env[63088]: DEBUG oslo_vmware.api [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]525d3e30-9b77-8b36-ef1f-bc2562d4bc2c, 'name': SearchDatastore_Task, 'duration_secs': 0.009239} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.551078] env[63088]: DEBUG oslo_concurrency.lockutils [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 926.551346] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] 72c07d1e-cbb6-4875-9b7f-2649aa25eee5/72c07d1e-cbb6-4875-9b7f-2649aa25eee5.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 926.551591] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a759b5b2-cdff-4e22-9b86-00d2159931ba {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.554639] env[63088]: DEBUG nova.network.neutron [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 926.558510] env[63088]: DEBUG oslo_vmware.api [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 926.558510] env[63088]: value = "task-1285206" [ 926.558510] env[63088]: _type = "Task" [ 926.558510] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.566319] env[63088]: DEBUG oslo_vmware.api [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285206, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.687178] env[63088]: DEBUG nova.network.neutron [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Updating instance_info_cache with network_info: [{"id": "81fc2dda-617e-4e4c-aa91-7e7020003995", "address": "fa:16:3e:8c:09:4d", "network": {"id": "1174a284-d3ca-4f9e-aa81-13ee9a693e55", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1994276040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa3d24a1a6c0430985fd80365d986ee1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap81fc2dda-61", "ovs_interfaceid": "81fc2dda-617e-4e4c-aa91-7e7020003995", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 926.773724] env[63088]: DEBUG oslo_concurrency.lockutils [None req-cd65c449-e671-45aa-a201-5e68b37dd782 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.743s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.776642] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4acba9f7-d735-4a08-90c3-58649f241640 tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.346s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.776642] env[63088]: DEBUG nova.objects.instance [None req-4acba9f7-d735-4a08-90c3-58649f241640 tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Lazy-loading 'resources' on Instance uuid bebc3318-24bb-4a37-8b23-66a12a7f7fd2 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 926.794416] env[63088]: INFO nova.scheduler.client.report [None req-cd65c449-e671-45aa-a201-5e68b37dd782 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Deleted allocations for instance 2964db35-7357-40a7-b4e6-7e2595549f5b [ 926.811355] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5c7fe3f5-a517-4b12-9489-f6aa83255a48 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Acquiring lock "dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.811656] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5c7fe3f5-a517-4b12-9489-f6aa83255a48 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Lock "dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.811876] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5c7fe3f5-a517-4b12-9489-f6aa83255a48 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Acquiring lock "dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.812522] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5c7fe3f5-a517-4b12-9489-f6aa83255a48 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Lock "dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.812713] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5c7fe3f5-a517-4b12-9489-f6aa83255a48 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Lock "dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.814952] env[63088]: INFO nova.compute.manager [None req-5c7fe3f5-a517-4b12-9489-f6aa83255a48 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Terminating instance [ 926.817054] env[63088]: DEBUG nova.compute.manager [None req-5c7fe3f5-a517-4b12-9489-f6aa83255a48 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 926.817230] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-5c7fe3f5-a517-4b12-9489-f6aa83255a48 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 926.818159] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8958ebb8-2a54-4f8c-83da-390d3e5b28ae {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.826604] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c7fe3f5-a517-4b12-9489-f6aa83255a48 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 926.826739] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0c0538b2-46cd-4366-af12-edc37feae2b8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.834184] env[63088]: DEBUG oslo_vmware.api [None req-5c7fe3f5-a517-4b12-9489-f6aa83255a48 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Waiting for the task: (returnval){ [ 926.834184] env[63088]: value = "task-1285207" [ 926.834184] env[63088]: _type = "Task" [ 926.834184] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.842400] env[63088]: DEBUG oslo_vmware.api [None req-5c7fe3f5-a517-4b12-9489-f6aa83255a48 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Task: {'id': task-1285207, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.882248] env[63088]: DEBUG oslo_concurrency.lockutils [None req-dcb8cf57-4815-4af4-b251-04f53799c533 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquiring lock "refresh_cache-7b6aadb7-e34b-42b7-b69f-370434f5b665" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 926.882581] env[63088]: DEBUG oslo_concurrency.lockutils [None req-dcb8cf57-4815-4af4-b251-04f53799c533 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquired lock "refresh_cache-7b6aadb7-e34b-42b7-b69f-370434f5b665" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 926.882744] env[63088]: DEBUG nova.network.neutron [None req-dcb8cf57-4815-4af4-b251-04f53799c533 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 926.928923] env[63088]: DEBUG oslo_concurrency.lockutils [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquiring lock "7cd45e52-c6be-4ec6-9710-fa99e29cf872" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.929345] env[63088]: DEBUG oslo_concurrency.lockutils [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lock "7cd45e52-c6be-4ec6-9710-fa99e29cf872" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.068421] env[63088]: DEBUG oslo_vmware.api [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285206, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.454451} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.068759] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] 72c07d1e-cbb6-4875-9b7f-2649aa25eee5/72c07d1e-cbb6-4875-9b7f-2649aa25eee5.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 927.069055] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] Extending root virtual disk to 1048576 {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 927.069350] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bb8546d1-f5db-4f89-9d2f-78cc5484a3f0 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.076024] env[63088]: DEBUG oslo_vmware.api [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 927.076024] env[63088]: value = "task-1285208" [ 927.076024] env[63088]: _type = "Task" [ 927.076024] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.084025] env[63088]: DEBUG oslo_vmware.api [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285208, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.189769] env[63088]: DEBUG nova.compute.manager [req-b0add4d2-4baa-4b54-9a3a-33f9c5776da1 req-f97fb0c7-3b7f-4ae9-8e41-e6382800675b service nova] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Received event network-vif-plugged-81fc2dda-617e-4e4c-aa91-7e7020003995 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 927.190031] env[63088]: DEBUG oslo_concurrency.lockutils [req-b0add4d2-4baa-4b54-9a3a-33f9c5776da1 req-f97fb0c7-3b7f-4ae9-8e41-e6382800675b service nova] Acquiring lock "f582da15-dfc3-45e1-a995-9dd0c9533869-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.190265] env[63088]: DEBUG oslo_concurrency.lockutils [req-b0add4d2-4baa-4b54-9a3a-33f9c5776da1 req-f97fb0c7-3b7f-4ae9-8e41-e6382800675b service nova] Lock "f582da15-dfc3-45e1-a995-9dd0c9533869-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.190442] env[63088]: DEBUG oslo_concurrency.lockutils [req-b0add4d2-4baa-4b54-9a3a-33f9c5776da1 req-f97fb0c7-3b7f-4ae9-8e41-e6382800675b service nova] Lock "f582da15-dfc3-45e1-a995-9dd0c9533869-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.190617] env[63088]: DEBUG nova.compute.manager [req-b0add4d2-4baa-4b54-9a3a-33f9c5776da1 req-f97fb0c7-3b7f-4ae9-8e41-e6382800675b service nova] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] No waiting events found dispatching network-vif-plugged-81fc2dda-617e-4e4c-aa91-7e7020003995 {{(pid=63088) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 927.190784] env[63088]: WARNING nova.compute.manager [req-b0add4d2-4baa-4b54-9a3a-33f9c5776da1 req-f97fb0c7-3b7f-4ae9-8e41-e6382800675b service nova] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Received unexpected event network-vif-plugged-81fc2dda-617e-4e4c-aa91-7e7020003995 for instance with vm_state building and task_state spawning. [ 927.190948] env[63088]: DEBUG nova.compute.manager [req-b0add4d2-4baa-4b54-9a3a-33f9c5776da1 req-f97fb0c7-3b7f-4ae9-8e41-e6382800675b service nova] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Received event network-changed-81fc2dda-617e-4e4c-aa91-7e7020003995 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 927.191356] env[63088]: DEBUG nova.compute.manager [req-b0add4d2-4baa-4b54-9a3a-33f9c5776da1 req-f97fb0c7-3b7f-4ae9-8e41-e6382800675b service nova] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Refreshing instance network info cache due to event network-changed-81fc2dda-617e-4e4c-aa91-7e7020003995. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 927.191499] env[63088]: DEBUG oslo_concurrency.lockutils [req-b0add4d2-4baa-4b54-9a3a-33f9c5776da1 req-f97fb0c7-3b7f-4ae9-8e41-e6382800675b service nova] Acquiring lock "refresh_cache-f582da15-dfc3-45e1-a995-9dd0c9533869" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 927.192774] env[63088]: DEBUG oslo_concurrency.lockutils [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Releasing lock "refresh_cache-f582da15-dfc3-45e1-a995-9dd0c9533869" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 927.192774] env[63088]: DEBUG nova.compute.manager [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Instance network_info: |[{"id": "81fc2dda-617e-4e4c-aa91-7e7020003995", "address": "fa:16:3e:8c:09:4d", "network": {"id": "1174a284-d3ca-4f9e-aa81-13ee9a693e55", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1994276040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa3d24a1a6c0430985fd80365d986ee1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap81fc2dda-61", "ovs_interfaceid": "81fc2dda-617e-4e4c-aa91-7e7020003995", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 927.192774] env[63088]: DEBUG oslo_concurrency.lockutils [req-b0add4d2-4baa-4b54-9a3a-33f9c5776da1 req-f97fb0c7-3b7f-4ae9-8e41-e6382800675b service nova] Acquired lock "refresh_cache-f582da15-dfc3-45e1-a995-9dd0c9533869" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 927.193042] env[63088]: DEBUG nova.network.neutron [req-b0add4d2-4baa-4b54-9a3a-33f9c5776da1 req-f97fb0c7-3b7f-4ae9-8e41-e6382800675b service nova] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Refreshing network info cache for port 81fc2dda-617e-4e4c-aa91-7e7020003995 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 927.194017] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8c:09:4d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '81fc2dda-617e-4e4c-aa91-7e7020003995', 'vif_model': 'vmxnet3'}] {{(pid=63088) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 927.202671] env[63088]: DEBUG oslo.service.loopingcall [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 927.205898] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Creating VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 927.206741] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-27e032e0-62d0-4b78-a98e-be41a35bdf90 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.227085] env[63088]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 927.227085] env[63088]: value = "task-1285209" [ 927.227085] env[63088]: _type = "Task" [ 927.227085] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.236693] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285209, 'name': CreateVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.307936] env[63088]: DEBUG oslo_concurrency.lockutils [None req-cd65c449-e671-45aa-a201-5e68b37dd782 tempest-ListServersNegativeTestJSON-1357025610 tempest-ListServersNegativeTestJSON-1357025610-project-member] Lock "2964db35-7357-40a7-b4e6-7e2595549f5b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.065s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.344928] env[63088]: DEBUG oslo_vmware.api [None req-5c7fe3f5-a517-4b12-9489-f6aa83255a48 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Task: {'id': task-1285207, 'name': PowerOffVM_Task, 'duration_secs': 0.286796} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.347329] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c7fe3f5-a517-4b12-9489-f6aa83255a48 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 927.347519] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-5c7fe3f5-a517-4b12-9489-f6aa83255a48 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Unregistering the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 927.348833] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-52a57caf-27b8-4819-b6a0-7e9f4433cf9f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.409502] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-5c7fe3f5-a517-4b12-9489-f6aa83255a48 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Unregistered the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 927.409785] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-5c7fe3f5-a517-4b12-9489-f6aa83255a48 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Deleting contents of the VM from datastore datastore2 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 927.409910] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c7fe3f5-a517-4b12-9489-f6aa83255a48 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Deleting the datastore file [datastore2] dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69 {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 927.410233] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b3b34cd5-c3f5-430c-bff8-c84a5916fba1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.416358] env[63088]: DEBUG oslo_vmware.api [None req-5c7fe3f5-a517-4b12-9489-f6aa83255a48 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Waiting for the task: (returnval){ [ 927.416358] env[63088]: value = "task-1285211" [ 927.416358] env[63088]: _type = "Task" [ 927.416358] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.429385] env[63088]: DEBUG oslo_vmware.api [None req-5c7fe3f5-a517-4b12-9489-f6aa83255a48 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Task: {'id': task-1285211, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.433347] env[63088]: DEBUG nova.compute.manager [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 7cd45e52-c6be-4ec6-9710-fa99e29cf872] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 927.494350] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4725c466-55b0-441a-b668-9cc083ef30fc {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.503621] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0aabfa0-b90c-47d1-8aec-c7db1a11321d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.544558] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be8426f7-79d4-4ccf-96f5-e00198387b95 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.552657] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd1c4453-d694-4b88-9aea-9d71d907f2ff {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.571051] env[63088]: DEBUG nova.compute.provider_tree [None req-4acba9f7-d735-4a08-90c3-58649f241640 tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 927.586966] env[63088]: DEBUG oslo_vmware.api [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285208, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.186222} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.587254] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] Extended root virtual disk {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 927.588038] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-357be2a1-3936-444d-9b27-304fbb64551c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.612261] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] Reconfiguring VM instance instance-00000053 to attach disk [datastore1] 72c07d1e-cbb6-4875-9b7f-2649aa25eee5/72c07d1e-cbb6-4875-9b7f-2649aa25eee5.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 927.613924] env[63088]: DEBUG nova.network.neutron [req-b0add4d2-4baa-4b54-9a3a-33f9c5776da1 req-f97fb0c7-3b7f-4ae9-8e41-e6382800675b service nova] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Updated VIF entry in instance network info cache for port 81fc2dda-617e-4e4c-aa91-7e7020003995. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 927.614388] env[63088]: DEBUG nova.network.neutron [req-b0add4d2-4baa-4b54-9a3a-33f9c5776da1 req-f97fb0c7-3b7f-4ae9-8e41-e6382800675b service nova] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Updating instance_info_cache with network_info: [{"id": "81fc2dda-617e-4e4c-aa91-7e7020003995", "address": "fa:16:3e:8c:09:4d", "network": {"id": "1174a284-d3ca-4f9e-aa81-13ee9a693e55", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1994276040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa3d24a1a6c0430985fd80365d986ee1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap81fc2dda-61", "ovs_interfaceid": "81fc2dda-617e-4e4c-aa91-7e7020003995", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 927.618283] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e409dd8a-5704-4709-8b9f-2f0d6694a028 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.639684] env[63088]: DEBUG oslo_vmware.api [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 927.639684] env[63088]: value = "task-1285212" [ 927.639684] env[63088]: _type = "Task" [ 927.639684] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.648922] env[63088]: DEBUG oslo_vmware.api [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285212, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.700073] env[63088]: DEBUG nova.network.neutron [None req-dcb8cf57-4815-4af4-b251-04f53799c533 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Updating instance_info_cache with network_info: [{"id": "fa4a91d1-df9a-4789-bc5c-a8b95457cd93", "address": "fa:16:3e:7e:c2:c7", "network": {"id": "dfcbe153-bc01-4362-9247-de9f0b1c847b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-841757397-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.152", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dace8b5181b84623b08f903d12dfd31e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c68b7663-4f0e-47f0-ac7f-40c6d952f7bb", "external-id": "nsx-vlan-transportzone-696", "segmentation_id": 696, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfa4a91d1-df", "ovs_interfaceid": "fa4a91d1-df9a-4789-bc5c-a8b95457cd93", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 927.737219] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285209, 'name': CreateVM_Task, 'duration_secs': 0.333697} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.737396] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Created VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 927.738157] env[63088]: DEBUG oslo_concurrency.lockutils [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 927.738342] env[63088]: DEBUG oslo_concurrency.lockutils [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 927.738661] env[63088]: DEBUG oslo_concurrency.lockutils [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 927.738932] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2d03e146-a00c-46e9-abcb-4db510aa83da {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.743504] env[63088]: DEBUG oslo_vmware.api [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Waiting for the task: (returnval){ [ 927.743504] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52da0741-81c7-b1f0-e845-2c58a96d3eb8" [ 927.743504] env[63088]: _type = "Task" [ 927.743504] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.751227] env[63088]: DEBUG oslo_vmware.api [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52da0741-81c7-b1f0-e845-2c58a96d3eb8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.932146] env[63088]: DEBUG oslo_vmware.api [None req-5c7fe3f5-a517-4b12-9489-f6aa83255a48 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Task: {'id': task-1285211, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.178314} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.932600] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c7fe3f5-a517-4b12-9489-f6aa83255a48 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 927.932863] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-5c7fe3f5-a517-4b12-9489-f6aa83255a48 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Deleted contents of the VM from datastore datastore2 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 927.933187] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-5c7fe3f5-a517-4b12-9489-f6aa83255a48 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 927.933365] env[63088]: INFO nova.compute.manager [None req-5c7fe3f5-a517-4b12-9489-f6aa83255a48 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Took 1.12 seconds to destroy the instance on the hypervisor. [ 927.933655] env[63088]: DEBUG oslo.service.loopingcall [None req-5c7fe3f5-a517-4b12-9489-f6aa83255a48 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 927.933875] env[63088]: DEBUG nova.compute.manager [-] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 927.934211] env[63088]: DEBUG nova.network.neutron [-] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 927.955900] env[63088]: DEBUG oslo_concurrency.lockutils [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 928.076432] env[63088]: DEBUG nova.scheduler.client.report [None req-4acba9f7-d735-4a08-90c3-58649f241640 tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 928.135323] env[63088]: DEBUG oslo_concurrency.lockutils [req-b0add4d2-4baa-4b54-9a3a-33f9c5776da1 req-f97fb0c7-3b7f-4ae9-8e41-e6382800675b service nova] Releasing lock "refresh_cache-f582da15-dfc3-45e1-a995-9dd0c9533869" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 928.155737] env[63088]: DEBUG oslo_vmware.api [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285212, 'name': ReconfigVM_Task, 'duration_secs': 0.270213} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.156086] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] Reconfigured VM instance instance-00000053 to attach disk [datastore1] 72c07d1e-cbb6-4875-9b7f-2649aa25eee5/72c07d1e-cbb6-4875-9b7f-2649aa25eee5.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 928.156799] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2a1aaa16-46c6-4ee4-8370-3ae8dd5abe05 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.164882] env[63088]: DEBUG oslo_vmware.api [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 928.164882] env[63088]: value = "task-1285213" [ 928.164882] env[63088]: _type = "Task" [ 928.164882] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.175015] env[63088]: DEBUG oslo_vmware.api [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285213, 'name': Rename_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.203201] env[63088]: DEBUG oslo_concurrency.lockutils [None req-dcb8cf57-4815-4af4-b251-04f53799c533 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Releasing lock "refresh_cache-7b6aadb7-e34b-42b7-b69f-370434f5b665" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 928.207855] env[63088]: DEBUG nova.compute.manager [None req-dcb8cf57-4815-4af4-b251-04f53799c533 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 928.208914] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efa7590b-bf71-49db-9955-f2e0804f2432 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.258586] env[63088]: DEBUG oslo_vmware.api [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52da0741-81c7-b1f0-e845-2c58a96d3eb8, 'name': SearchDatastore_Task, 'duration_secs': 0.010133} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.259045] env[63088]: DEBUG oslo_concurrency.lockutils [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 928.259445] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Processing image 1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 928.259811] env[63088]: DEBUG oslo_concurrency.lockutils [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 928.260014] env[63088]: DEBUG oslo_concurrency.lockutils [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 928.260277] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 928.261314] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ceb9e0a1-9ab8-4bab-bc7b-c1bd58adc321 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.270631] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 928.274022] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 928.274022] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e49d83ee-1a50-4eef-b166-e5be5e5cfb1c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.278432] env[63088]: DEBUG oslo_vmware.api [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Waiting for the task: (returnval){ [ 928.278432] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]526bf113-431b-acc9-b636-0c601d20b8d1" [ 928.278432] env[63088]: _type = "Task" [ 928.278432] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.287697] env[63088]: DEBUG oslo_vmware.api [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]526bf113-431b-acc9-b636-0c601d20b8d1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.583653] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4acba9f7-d735-4a08-90c3-58649f241640 tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.807s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.586152] env[63088]: DEBUG oslo_concurrency.lockutils [None req-901dfe38-2626-4a52-ab00-0a33f4bc1b1f tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.572s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 928.586454] env[63088]: DEBUG nova.objects.instance [None req-901dfe38-2626-4a52-ab00-0a33f4bc1b1f tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Lazy-loading 'resources' on Instance uuid 45513c6b-c000-4ee1-8893-4e084ffc22c5 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 928.617371] env[63088]: INFO nova.scheduler.client.report [None req-4acba9f7-d735-4a08-90c3-58649f241640 tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Deleted allocations for instance bebc3318-24bb-4a37-8b23-66a12a7f7fd2 [ 928.676156] env[63088]: DEBUG oslo_vmware.api [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285213, 'name': Rename_Task, 'duration_secs': 0.142478} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.676484] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 928.676755] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1c8cdded-00fe-4fb2-9915-83679a74e58a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.683291] env[63088]: DEBUG oslo_vmware.api [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 928.683291] env[63088]: value = "task-1285214" [ 928.683291] env[63088]: _type = "Task" [ 928.683291] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.691239] env[63088]: DEBUG oslo_vmware.api [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285214, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.791550] env[63088]: DEBUG oslo_vmware.api [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]526bf113-431b-acc9-b636-0c601d20b8d1, 'name': SearchDatastore_Task, 'duration_secs': 0.013701} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.794466] env[63088]: DEBUG nova.compute.manager [req-17dfa4a9-b3ed-45bb-989a-7637cd85179d req-57293853-3c3b-4931-ba23-e7d2d1d0c448 service nova] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Received event network-vif-deleted-6ce9925d-cf59-4de0-aece-da90fb4b9f36 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 928.794826] env[63088]: INFO nova.compute.manager [req-17dfa4a9-b3ed-45bb-989a-7637cd85179d req-57293853-3c3b-4931-ba23-e7d2d1d0c448 service nova] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Neutron deleted interface 6ce9925d-cf59-4de0-aece-da90fb4b9f36; detaching it from the instance and deleting it from the info cache [ 928.795063] env[63088]: DEBUG nova.network.neutron [req-17dfa4a9-b3ed-45bb-989a-7637cd85179d req-57293853-3c3b-4931-ba23-e7d2d1d0c448 service nova] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 928.796094] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bbb403ec-9ae6-466d-8e39-50df2393d861 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.804516] env[63088]: DEBUG oslo_vmware.api [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Waiting for the task: (returnval){ [ 928.804516] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52780726-3da8-755f-28eb-d4a1073ef7c0" [ 928.804516] env[63088]: _type = "Task" [ 928.804516] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.813532] env[63088]: DEBUG oslo_vmware.api [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52780726-3da8-755f-28eb-d4a1073ef7c0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.130069] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4acba9f7-d735-4a08-90c3-58649f241640 tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Lock "bebc3318-24bb-4a37-8b23-66a12a7f7fd2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.628s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.194857] env[63088]: DEBUG oslo_vmware.api [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285214, 'name': PowerOnVM_Task, 'duration_secs': 0.473567} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.195211] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 929.195428] env[63088]: INFO nova.compute.manager [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] Took 6.33 seconds to spawn the instance on the hypervisor. [ 929.196393] env[63088]: DEBUG nova.compute.manager [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 929.197217] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e07ee3c-7f55-41c7-8a14-060f60608f18 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.225763] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9ff6219-168e-4dea-823d-d5e37b4cbd3b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.233182] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-dcb8cf57-4815-4af4-b251-04f53799c533 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Doing hard reboot of VM {{(pid=63088) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1064}} [ 929.235457] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-064d2314-05d0-4162-a84e-1847be1ed8dc {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.241601] env[63088]: DEBUG oslo_vmware.api [None req-dcb8cf57-4815-4af4-b251-04f53799c533 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for the task: (returnval){ [ 929.241601] env[63088]: value = "task-1285215" [ 929.241601] env[63088]: _type = "Task" [ 929.241601] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.256843] env[63088]: DEBUG oslo_vmware.api [None req-dcb8cf57-4815-4af4-b251-04f53799c533 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285215, 'name': ResetVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.262764] env[63088]: DEBUG nova.network.neutron [-] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 929.290533] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6dff0d3-87b2-4d9d-9f3c-e75280959374 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.299071] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adb8f744-4d26-4594-b786-31c499bbccac {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.302724] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d07aa124-3b5b-4538-9e29-d8c97c5cf045 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.343702] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73258a91-0bac-47db-b11c-8bfc13b55e4e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.353569] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24bf11dd-3ba5-413d-badf-6c27c321fead {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.364468] env[63088]: DEBUG oslo_vmware.api [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52780726-3da8-755f-28eb-d4a1073ef7c0, 'name': SearchDatastore_Task, 'duration_secs': 0.015854} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.365228] env[63088]: DEBUG oslo_concurrency.lockutils [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 929.365519] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] f582da15-dfc3-45e1-a995-9dd0c9533869/f582da15-dfc3-45e1-a995-9dd0c9533869.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 929.366113] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-54f7169e-c174-4896-b57b-b925a321ae96 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.371758] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6dc03c2-1844-43f9-bc3e-294ea221ce67 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.385533] env[63088]: DEBUG nova.compute.manager [req-17dfa4a9-b3ed-45bb-989a-7637cd85179d req-57293853-3c3b-4931-ba23-e7d2d1d0c448 service nova] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Detach interface failed, port_id=6ce9925d-cf59-4de0-aece-da90fb4b9f36, reason: Instance dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69 could not be found. {{(pid=63088) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 929.386425] env[63088]: DEBUG oslo_vmware.api [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Waiting for the task: (returnval){ [ 929.386425] env[63088]: value = "task-1285216" [ 929.386425] env[63088]: _type = "Task" [ 929.386425] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.399213] env[63088]: DEBUG nova.compute.provider_tree [None req-901dfe38-2626-4a52-ab00-0a33f4bc1b1f tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 929.405536] env[63088]: DEBUG oslo_vmware.api [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285216, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.717837] env[63088]: INFO nova.compute.manager [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] Took 15.68 seconds to build instance. [ 929.756016] env[63088]: DEBUG oslo_vmware.api [None req-dcb8cf57-4815-4af4-b251-04f53799c533 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285215, 'name': ResetVM_Task, 'duration_secs': 0.090547} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.756390] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-dcb8cf57-4815-4af4-b251-04f53799c533 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Did hard reboot of VM {{(pid=63088) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1068}} [ 929.756808] env[63088]: DEBUG nova.compute.manager [None req-dcb8cf57-4815-4af4-b251-04f53799c533 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 929.757677] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3077bbcd-3b1c-4400-8c96-80ca01fa2269 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.768214] env[63088]: INFO nova.compute.manager [-] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Took 1.83 seconds to deallocate network for instance. [ 929.896453] env[63088]: DEBUG oslo_vmware.api [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285216, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.902786] env[63088]: DEBUG nova.scheduler.client.report [None req-901dfe38-2626-4a52-ab00-0a33f4bc1b1f tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 930.219522] env[63088]: DEBUG oslo_concurrency.lockutils [None req-ad89730a-a71c-4d1e-86dd-d723ceff8aa2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "72c07d1e-cbb6-4875-9b7f-2649aa25eee5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.188s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.271941] env[63088]: DEBUG oslo_concurrency.lockutils [None req-dcb8cf57-4815-4af4-b251-04f53799c533 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lock "7b6aadb7-e34b-42b7-b69f-370434f5b665" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 3.914s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.275176] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5c7fe3f5-a517-4b12-9489-f6aa83255a48 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 930.399867] env[63088]: DEBUG oslo_vmware.api [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285216, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.817782} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.400233] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] f582da15-dfc3-45e1-a995-9dd0c9533869/f582da15-dfc3-45e1-a995-9dd0c9533869.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 930.400522] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Extending root virtual disk to 1048576 {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 930.400841] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-40c689f0-c119-4c5d-a5bc-0d006f3fa7f2 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.412128] env[63088]: DEBUG oslo_concurrency.lockutils [None req-901dfe38-2626-4a52-ab00-0a33f4bc1b1f tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.822s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.414868] env[63088]: DEBUG oslo_concurrency.lockutils [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.578s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 930.419295] env[63088]: INFO nova.compute.claims [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 79c4d6e8-9999-4b0c-98d3-bbfd4d61e660] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 930.427081] env[63088]: DEBUG oslo_vmware.api [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Waiting for the task: (returnval){ [ 930.427081] env[63088]: value = "task-1285217" [ 930.427081] env[63088]: _type = "Task" [ 930.427081] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.438580] env[63088]: DEBUG oslo_vmware.api [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285217, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.443732] env[63088]: INFO nova.scheduler.client.report [None req-901dfe38-2626-4a52-ab00-0a33f4bc1b1f tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Deleted allocations for instance 45513c6b-c000-4ee1-8893-4e084ffc22c5 [ 930.938221] env[63088]: DEBUG oslo_vmware.api [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285217, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063523} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.938221] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Extended root virtual disk {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 930.938489] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-249d98d2-32a1-48d6-855f-4e19057f163d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.964369] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Reconfiguring VM instance instance-00000052 to attach disk [datastore1] f582da15-dfc3-45e1-a995-9dd0c9533869/f582da15-dfc3-45e1-a995-9dd0c9533869.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 930.965742] env[63088]: DEBUG oslo_concurrency.lockutils [None req-901dfe38-2626-4a52-ab00-0a33f4bc1b1f tempest-MultipleCreateTestJSON-170902413 tempest-MultipleCreateTestJSON-170902413-project-member] Lock "45513c6b-c000-4ee1-8893-4e084ffc22c5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.358s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.971736] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8c7151c4-201a-4e39-9963-84502c5df17f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.003034] env[63088]: DEBUG oslo_vmware.api [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Waiting for the task: (returnval){ [ 931.003034] env[63088]: value = "task-1285218" [ 931.003034] env[63088]: _type = "Task" [ 931.003034] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.012800] env[63088]: DEBUG oslo_vmware.api [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285218, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.498751] env[63088]: DEBUG oslo_concurrency.lockutils [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquiring lock "72c07d1e-cbb6-4875-9b7f-2649aa25eee5" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 931.499056] env[63088]: DEBUG oslo_concurrency.lockutils [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "72c07d1e-cbb6-4875-9b7f-2649aa25eee5" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 931.499247] env[63088]: INFO nova.compute.manager [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] Shelving [ 931.516359] env[63088]: DEBUG oslo_vmware.api [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285218, 'name': ReconfigVM_Task, 'duration_secs': 0.293278} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.516638] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Reconfigured VM instance instance-00000052 to attach disk [datastore1] f582da15-dfc3-45e1-a995-9dd0c9533869/f582da15-dfc3-45e1-a995-9dd0c9533869.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 931.518105] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-67feb6da-8df4-48ab-a1f3-518bf714eada {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.528904] env[63088]: DEBUG oslo_vmware.api [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Waiting for the task: (returnval){ [ 931.528904] env[63088]: value = "task-1285219" [ 931.528904] env[63088]: _type = "Task" [ 931.528904] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.542517] env[63088]: DEBUG oslo_vmware.api [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285219, 'name': Rename_Task} progress is 6%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.632201] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8493ef7-147b-4a47-bb69-1e18464123b5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.639615] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1e4bb77-2e02-4a88-9a80-c1a61b1b6be9 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.675445] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0a4b547-590f-4d17-a452-7361969779f1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.683556] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-249a5e4d-bbbc-4c62-ab7a-6384f6e32cad {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.698996] env[63088]: DEBUG nova.compute.provider_tree [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 932.010480] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 932.010796] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d043df11-6b8c-40b6-9ca4-fdd393ec8c89 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.018747] env[63088]: DEBUG oslo_vmware.api [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 932.018747] env[63088]: value = "task-1285220" [ 932.018747] env[63088]: _type = "Task" [ 932.018747] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.028138] env[63088]: DEBUG oslo_vmware.api [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285220, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.037996] env[63088]: DEBUG oslo_vmware.api [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285219, 'name': Rename_Task, 'duration_secs': 0.137215} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.038352] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 932.038619] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d9b5cd8e-a809-40bb-b3ac-fc2ed3e9ba7e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.045716] env[63088]: DEBUG oslo_vmware.api [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Waiting for the task: (returnval){ [ 932.045716] env[63088]: value = "task-1285221" [ 932.045716] env[63088]: _type = "Task" [ 932.045716] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.056572] env[63088]: DEBUG oslo_vmware.api [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285221, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.204309] env[63088]: DEBUG nova.scheduler.client.report [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 932.530015] env[63088]: DEBUG oslo_vmware.api [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285220, 'name': PowerOffVM_Task, 'duration_secs': 0.278898} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.530015] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 932.530015] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cf9bf64-511c-42b6-9cf5-aaeff35fe35c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.553209] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acc5c746-04f8-4643-8b43-f7cb6546c3a3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.570161] env[63088]: DEBUG oslo_vmware.api [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285221, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.708539] env[63088]: DEBUG oslo_concurrency.lockutils [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.294s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.709118] env[63088]: DEBUG nova.compute.manager [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 79c4d6e8-9999-4b0c-98d3-bbfd4d61e660] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 932.715362] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.675s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.717272] env[63088]: INFO nova.compute.claims [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2ad33254-8030-4454-b023-3359071f85de] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 933.060074] env[63088]: DEBUG oslo_vmware.api [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285221, 'name': PowerOnVM_Task, 'duration_secs': 0.604363} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.060452] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 933.060720] env[63088]: INFO nova.compute.manager [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Took 7.89 seconds to spawn the instance on the hypervisor. [ 933.060953] env[63088]: DEBUG nova.compute.manager [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 933.061958] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d577cd8a-0cb6-40c2-9a37-3cca0efa9535 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.073597] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] Creating Snapshot of the VM instance {{(pid=63088) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 933.073915] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-b90ecfbb-556c-4ed0-9a23-b9bcbfff355f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.081975] env[63088]: DEBUG oslo_vmware.api [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 933.081975] env[63088]: value = "task-1285222" [ 933.081975] env[63088]: _type = "Task" [ 933.081975] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.090254] env[63088]: DEBUG oslo_vmware.api [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285222, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.171340] env[63088]: DEBUG oslo_concurrency.lockutils [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Acquiring lock "8205f9e1-effd-442d-990b-bf5a2a9e6cdd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.171588] env[63088]: DEBUG oslo_concurrency.lockutils [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Lock "8205f9e1-effd-442d-990b-bf5a2a9e6cdd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.222216] env[63088]: DEBUG nova.compute.utils [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 933.225833] env[63088]: DEBUG nova.compute.manager [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 79c4d6e8-9999-4b0c-98d3-bbfd4d61e660] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 933.225833] env[63088]: DEBUG nova.network.neutron [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 79c4d6e8-9999-4b0c-98d3-bbfd4d61e660] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 933.288054] env[63088]: DEBUG nova.policy [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a6f91fc8922c42f985f6f6338dab0708', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7bff1382b9694df08133c88a5fe783a5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 933.593677] env[63088]: INFO nova.compute.manager [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Took 19.52 seconds to build instance. [ 933.601910] env[63088]: DEBUG oslo_vmware.api [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285222, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.631924] env[63088]: DEBUG nova.network.neutron [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 79c4d6e8-9999-4b0c-98d3-bbfd4d61e660] Successfully created port: 933dd0ba-3083-4fec-bc5f-1270e8761e92 {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 933.674409] env[63088]: DEBUG nova.compute.manager [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] [instance: 8205f9e1-effd-442d-990b-bf5a2a9e6cdd] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 933.728609] env[63088]: DEBUG nova.compute.manager [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 79c4d6e8-9999-4b0c-98d3-bbfd4d61e660] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 933.971252] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d35c09c4-a366-493e-a83a-7b63ceac8121 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.980607] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e674d5e-e1ff-4d8a-afa3-6e352c77e159 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.015689] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-773353bb-14ee-4295-9387-c3d02da4aceb {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.024399] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d74cbee-d6cf-4504-8a75-2aa02ff7801d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.038577] env[63088]: DEBUG nova.compute.provider_tree [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 934.102225] env[63088]: DEBUG oslo_concurrency.lockutils [None req-340c3a25-9272-4c4b-8f6d-f7faef4f0296 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Lock "f582da15-dfc3-45e1-a995-9dd0c9533869" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.031s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 934.102570] env[63088]: DEBUG oslo_vmware.api [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285222, 'name': CreateSnapshot_Task, 'duration_secs': 0.776023} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.105021] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] Created Snapshot of the VM instance {{(pid=63088) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 934.105021] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50fdcaa3-524e-46ec-aca0-0815cd5c198f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.195788] env[63088]: DEBUG oslo_concurrency.lockutils [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 934.542629] env[63088]: DEBUG nova.scheduler.client.report [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 934.637523] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] Creating linked-clone VM from snapshot {{(pid=63088) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 934.637523] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-bc649f4d-a2de-468f-b238-8dff9aaa5b9b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.648778] env[63088]: DEBUG oslo_vmware.api [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 934.648778] env[63088]: value = "task-1285223" [ 934.648778] env[63088]: _type = "Task" [ 934.648778] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.663601] env[63088]: DEBUG oslo_vmware.api [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285223, 'name': CloneVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.666570] env[63088]: DEBUG nova.compute.manager [req-a3037fef-c7d6-4707-a996-4fd05237cd0a req-92ccf946-1c3d-40cd-8e0a-00a666f85d0a service nova] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Received event network-changed-81fc2dda-617e-4e4c-aa91-7e7020003995 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 934.666885] env[63088]: DEBUG nova.compute.manager [req-a3037fef-c7d6-4707-a996-4fd05237cd0a req-92ccf946-1c3d-40cd-8e0a-00a666f85d0a service nova] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Refreshing instance network info cache due to event network-changed-81fc2dda-617e-4e4c-aa91-7e7020003995. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 934.667249] env[63088]: DEBUG oslo_concurrency.lockutils [req-a3037fef-c7d6-4707-a996-4fd05237cd0a req-92ccf946-1c3d-40cd-8e0a-00a666f85d0a service nova] Acquiring lock "refresh_cache-f582da15-dfc3-45e1-a995-9dd0c9533869" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 934.667526] env[63088]: DEBUG oslo_concurrency.lockutils [req-a3037fef-c7d6-4707-a996-4fd05237cd0a req-92ccf946-1c3d-40cd-8e0a-00a666f85d0a service nova] Acquired lock "refresh_cache-f582da15-dfc3-45e1-a995-9dd0c9533869" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 934.668886] env[63088]: DEBUG nova.network.neutron [req-a3037fef-c7d6-4707-a996-4fd05237cd0a req-92ccf946-1c3d-40cd-8e0a-00a666f85d0a service nova] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Refreshing network info cache for port 81fc2dda-617e-4e4c-aa91-7e7020003995 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 934.742662] env[63088]: DEBUG nova.compute.manager [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 79c4d6e8-9999-4b0c-98d3-bbfd4d61e660] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 934.770207] env[63088]: DEBUG nova.virt.hardware [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='6b7f7990180cce41fea32db309f0c9d1',container_format='bare',created_at=2024-10-15T12:20:09Z,direct_url=,disk_format='vmdk',id=c89225f6-2d34-4988-ae47-d3d6903890da,min_disk=1,min_ram=0,name='tempest-test-snap-839309440',owner='7bff1382b9694df08133c88a5fe783a5',properties=ImageMetaProps,protected=,size=21334016,status='active',tags=,updated_at=2024-10-15T12:20:26Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 934.770460] env[63088]: DEBUG nova.virt.hardware [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 934.770624] env[63088]: DEBUG nova.virt.hardware [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 934.770813] env[63088]: DEBUG nova.virt.hardware [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 934.771036] env[63088]: DEBUG nova.virt.hardware [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 934.771233] env[63088]: DEBUG nova.virt.hardware [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 934.771465] env[63088]: DEBUG nova.virt.hardware [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 934.771635] env[63088]: DEBUG nova.virt.hardware [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 934.771808] env[63088]: DEBUG nova.virt.hardware [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 934.771976] env[63088]: DEBUG nova.virt.hardware [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 934.772173] env[63088]: DEBUG nova.virt.hardware [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 934.773317] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a664cf42-ca51-49d5-919d-541cecb7500e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.782550] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adcef923-83f9-44ce-919f-8734006465b1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.053495] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.338s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.054152] env[63088]: DEBUG nova.compute.manager [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2ad33254-8030-4454-b023-3359071f85de] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 935.057081] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5a7bffe6-5d81-4eff-949f-17e8fd61a581 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.523s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.057269] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5a7bffe6-5d81-4eff-949f-17e8fd61a581 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.059864] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3d5df639-f016-4d69-8403-3ba3b2d15e1d tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.874s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.060206] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3d5df639-f016-4d69-8403-3ba3b2d15e1d tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.062218] env[63088]: DEBUG oslo_concurrency.lockutils [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.106s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.063851] env[63088]: INFO nova.compute.claims [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 7cd45e52-c6be-4ec6-9710-fa99e29cf872] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 935.107622] env[63088]: INFO nova.scheduler.client.report [None req-3d5df639-f016-4d69-8403-3ba3b2d15e1d tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Deleted allocations for instance e6b0ce08-d67f-458f-92f5-1e904d03a5ef [ 935.113157] env[63088]: INFO nova.scheduler.client.report [None req-5a7bffe6-5d81-4eff-949f-17e8fd61a581 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Deleted allocations for instance 2269e520-08b6-433f-8fe9-7b84d91e02d7 [ 935.162275] env[63088]: DEBUG oslo_vmware.api [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285223, 'name': CloneVM_Task} progress is 94%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.252175] env[63088]: DEBUG nova.compute.manager [req-9b88389b-f331-46cb-b243-54cb7f4fdb2a req-fcbb1245-b414-43a1-897b-f1088ff8c218 service nova] [instance: 79c4d6e8-9999-4b0c-98d3-bbfd4d61e660] Received event network-vif-plugged-933dd0ba-3083-4fec-bc5f-1270e8761e92 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 935.252175] env[63088]: DEBUG oslo_concurrency.lockutils [req-9b88389b-f331-46cb-b243-54cb7f4fdb2a req-fcbb1245-b414-43a1-897b-f1088ff8c218 service nova] Acquiring lock "79c4d6e8-9999-4b0c-98d3-bbfd4d61e660-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.252509] env[63088]: DEBUG oslo_concurrency.lockutils [req-9b88389b-f331-46cb-b243-54cb7f4fdb2a req-fcbb1245-b414-43a1-897b-f1088ff8c218 service nova] Lock "79c4d6e8-9999-4b0c-98d3-bbfd4d61e660-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.252509] env[63088]: DEBUG oslo_concurrency.lockutils [req-9b88389b-f331-46cb-b243-54cb7f4fdb2a req-fcbb1245-b414-43a1-897b-f1088ff8c218 service nova] Lock "79c4d6e8-9999-4b0c-98d3-bbfd4d61e660-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.252645] env[63088]: DEBUG nova.compute.manager [req-9b88389b-f331-46cb-b243-54cb7f4fdb2a req-fcbb1245-b414-43a1-897b-f1088ff8c218 service nova] [instance: 79c4d6e8-9999-4b0c-98d3-bbfd4d61e660] No waiting events found dispatching network-vif-plugged-933dd0ba-3083-4fec-bc5f-1270e8761e92 {{(pid=63088) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 935.252803] env[63088]: WARNING nova.compute.manager [req-9b88389b-f331-46cb-b243-54cb7f4fdb2a req-fcbb1245-b414-43a1-897b-f1088ff8c218 service nova] [instance: 79c4d6e8-9999-4b0c-98d3-bbfd4d61e660] Received unexpected event network-vif-plugged-933dd0ba-3083-4fec-bc5f-1270e8761e92 for instance with vm_state building and task_state spawning. [ 935.408021] env[63088]: DEBUG nova.network.neutron [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 79c4d6e8-9999-4b0c-98d3-bbfd4d61e660] Successfully updated port: 933dd0ba-3083-4fec-bc5f-1270e8761e92 {{(pid=63088) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 935.465446] env[63088]: DEBUG nova.network.neutron [req-a3037fef-c7d6-4707-a996-4fd05237cd0a req-92ccf946-1c3d-40cd-8e0a-00a666f85d0a service nova] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Updated VIF entry in instance network info cache for port 81fc2dda-617e-4e4c-aa91-7e7020003995. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 935.465840] env[63088]: DEBUG nova.network.neutron [req-a3037fef-c7d6-4707-a996-4fd05237cd0a req-92ccf946-1c3d-40cd-8e0a-00a666f85d0a service nova] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Updating instance_info_cache with network_info: [{"id": "81fc2dda-617e-4e4c-aa91-7e7020003995", "address": "fa:16:3e:8c:09:4d", "network": {"id": "1174a284-d3ca-4f9e-aa81-13ee9a693e55", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1994276040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.150", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa3d24a1a6c0430985fd80365d986ee1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap81fc2dda-61", "ovs_interfaceid": "81fc2dda-617e-4e4c-aa91-7e7020003995", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 935.572623] env[63088]: DEBUG nova.compute.utils [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 935.577394] env[63088]: DEBUG nova.compute.manager [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2ad33254-8030-4454-b023-3359071f85de] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 935.577583] env[63088]: DEBUG nova.network.neutron [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2ad33254-8030-4454-b023-3359071f85de] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 935.622900] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3d5df639-f016-4d69-8403-3ba3b2d15e1d tempest-MigrationsAdminTest-1958059272 tempest-MigrationsAdminTest-1958059272-project-member] Lock "e6b0ce08-d67f-458f-92f5-1e904d03a5ef" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.435s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.629331] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5a7bffe6-5d81-4eff-949f-17e8fd61a581 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Lock "2269e520-08b6-433f-8fe9-7b84d91e02d7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.145s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.636667] env[63088]: DEBUG nova.policy [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4424adc5bbe34a91b669c3a907882644', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f33f2701fad94864a8c406a404bc0a42', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 935.662642] env[63088]: DEBUG oslo_vmware.api [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285223, 'name': CloneVM_Task} progress is 95%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.796996] env[63088]: DEBUG oslo_concurrency.lockutils [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquiring lock "d6fef3ef-fec8-4929-b9b8-5e63306aa51d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.797334] env[63088]: DEBUG oslo_concurrency.lockutils [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Lock "d6fef3ef-fec8-4929-b9b8-5e63306aa51d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.911436] env[63088]: DEBUG oslo_concurrency.lockutils [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Acquiring lock "refresh_cache-79c4d6e8-9999-4b0c-98d3-bbfd4d61e660" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 935.911588] env[63088]: DEBUG oslo_concurrency.lockutils [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Acquired lock "refresh_cache-79c4d6e8-9999-4b0c-98d3-bbfd4d61e660" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 935.911738] env[63088]: DEBUG nova.network.neutron [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 79c4d6e8-9999-4b0c-98d3-bbfd4d61e660] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 935.968855] env[63088]: DEBUG oslo_concurrency.lockutils [req-a3037fef-c7d6-4707-a996-4fd05237cd0a req-92ccf946-1c3d-40cd-8e0a-00a666f85d0a service nova] Releasing lock "refresh_cache-f582da15-dfc3-45e1-a995-9dd0c9533869" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 936.002199] env[63088]: DEBUG nova.network.neutron [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2ad33254-8030-4454-b023-3359071f85de] Successfully created port: 1d2bee7e-bbc7-4b17-976b-47e7bbb268af {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 936.080070] env[63088]: DEBUG nova.compute.manager [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2ad33254-8030-4454-b023-3359071f85de] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 936.169054] env[63088]: DEBUG oslo_vmware.api [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285223, 'name': CloneVM_Task, 'duration_secs': 1.342359} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.169238] env[63088]: INFO nova.virt.vmwareapi.vmops [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] Created linked-clone VM from snapshot [ 936.170054] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ad472d3-38f0-40c6-9bc3-55d624425e94 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.183463] env[63088]: DEBUG nova.virt.vmwareapi.images [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] Uploading image e0b4e03a-baf3-4fa2-ba4f-555963d990d6 {{(pid=63088) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 936.217497] env[63088]: DEBUG oslo_vmware.rw_handles [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 936.217497] env[63088]: value = "vm-275912" [ 936.217497] env[63088]: _type = "VirtualMachine" [ 936.217497] env[63088]: }. {{(pid=63088) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 936.217818] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-d7dc85dd-a3a4-4de0-9ecc-28af0f7aa301 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.228708] env[63088]: DEBUG oslo_vmware.rw_handles [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lease: (returnval){ [ 936.228708] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]5271cd55-acda-322e-29e5-831b64a6fefc" [ 936.228708] env[63088]: _type = "HttpNfcLease" [ 936.228708] env[63088]: } obtained for exporting VM: (result){ [ 936.228708] env[63088]: value = "vm-275912" [ 936.228708] env[63088]: _type = "VirtualMachine" [ 936.228708] env[63088]: }. {{(pid=63088) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 936.228964] env[63088]: DEBUG oslo_vmware.api [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the lease: (returnval){ [ 936.228964] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]5271cd55-acda-322e-29e5-831b64a6fefc" [ 936.228964] env[63088]: _type = "HttpNfcLease" [ 936.228964] env[63088]: } to be ready. {{(pid=63088) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 936.237061] env[63088]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 936.237061] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]5271cd55-acda-322e-29e5-831b64a6fefc" [ 936.237061] env[63088]: _type = "HttpNfcLease" [ 936.237061] env[63088]: } is initializing. {{(pid=63088) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 936.301395] env[63088]: DEBUG nova.compute.manager [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 936.322747] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1e36874-502c-4f63-aa35-a012d3e812ae {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.331020] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d6c7ee8-bdee-4892-ac4f-4326fe76784a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.367759] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a322f896-efdd-4777-a741-66588f0d73bb {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.376406] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed91763d-64d0-4340-852d-e2299c1661cd {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.392532] env[63088]: DEBUG nova.compute.provider_tree [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 936.458200] env[63088]: DEBUG nova.network.neutron [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 79c4d6e8-9999-4b0c-98d3-bbfd4d61e660] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 936.651013] env[63088]: DEBUG nova.network.neutron [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 79c4d6e8-9999-4b0c-98d3-bbfd4d61e660] Updating instance_info_cache with network_info: [{"id": "933dd0ba-3083-4fec-bc5f-1270e8761e92", "address": "fa:16:3e:4f:7c:f1", "network": {"id": "2bbaf351-fe63-46ba-ba7e-0f432c0a1bfb", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1438884226-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7bff1382b9694df08133c88a5fe783a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cca1f087-01e1-49ca-831b-5c51478a5d60", "external-id": "nsx-vlan-transportzone-439", "segmentation_id": 439, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap933dd0ba-30", "ovs_interfaceid": "933dd0ba-3083-4fec-bc5f-1270e8761e92", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 936.738431] env[63088]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 936.738431] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]5271cd55-acda-322e-29e5-831b64a6fefc" [ 936.738431] env[63088]: _type = "HttpNfcLease" [ 936.738431] env[63088]: } is ready. {{(pid=63088) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 936.738831] env[63088]: DEBUG oslo_vmware.rw_handles [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 936.738831] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]5271cd55-acda-322e-29e5-831b64a6fefc" [ 936.738831] env[63088]: _type = "HttpNfcLease" [ 936.738831] env[63088]: }. {{(pid=63088) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 936.739897] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adc9866d-8bdb-4cc1-a42e-9b0e178dce22 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.747546] env[63088]: DEBUG oslo_vmware.rw_handles [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526bdc2d-92d1-17cb-9ddc-6c73bb412339/disk-0.vmdk from lease info. {{(pid=63088) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 936.747546] env[63088]: DEBUG oslo_vmware.rw_handles [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526bdc2d-92d1-17cb-9ddc-6c73bb412339/disk-0.vmdk for reading. {{(pid=63088) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 936.824562] env[63088]: DEBUG oslo_concurrency.lockutils [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 936.835101] env[63088]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-142fd5b9-7e8a-44f5-a39e-dea509a8b16d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.896447] env[63088]: DEBUG nova.scheduler.client.report [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 937.096084] env[63088]: DEBUG nova.compute.manager [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2ad33254-8030-4454-b023-3359071f85de] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 937.127206] env[63088]: DEBUG nova.virt.hardware [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 937.127708] env[63088]: DEBUG nova.virt.hardware [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 937.129860] env[63088]: DEBUG nova.virt.hardware [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 937.129860] env[63088]: DEBUG nova.virt.hardware [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 937.129860] env[63088]: DEBUG nova.virt.hardware [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 937.129860] env[63088]: DEBUG nova.virt.hardware [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 937.129860] env[63088]: DEBUG nova.virt.hardware [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 937.129860] env[63088]: DEBUG nova.virt.hardware [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 937.129860] env[63088]: DEBUG nova.virt.hardware [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 937.129860] env[63088]: DEBUG nova.virt.hardware [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 937.129860] env[63088]: DEBUG nova.virt.hardware [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 937.131377] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-740593df-46c7-476a-af37-ceabdb7908b1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.141643] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-378968cf-9f25-4d9d-9384-31dfbf27d62e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.157717] env[63088]: DEBUG oslo_concurrency.lockutils [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Releasing lock "refresh_cache-79c4d6e8-9999-4b0c-98d3-bbfd4d61e660" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 937.158081] env[63088]: DEBUG nova.compute.manager [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 79c4d6e8-9999-4b0c-98d3-bbfd4d61e660] Instance network_info: |[{"id": "933dd0ba-3083-4fec-bc5f-1270e8761e92", "address": "fa:16:3e:4f:7c:f1", "network": {"id": "2bbaf351-fe63-46ba-ba7e-0f432c0a1bfb", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1438884226-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7bff1382b9694df08133c88a5fe783a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cca1f087-01e1-49ca-831b-5c51478a5d60", "external-id": "nsx-vlan-transportzone-439", "segmentation_id": 439, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap933dd0ba-30", "ovs_interfaceid": "933dd0ba-3083-4fec-bc5f-1270e8761e92", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 937.158797] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 79c4d6e8-9999-4b0c-98d3-bbfd4d61e660] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4f:7c:f1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cca1f087-01e1-49ca-831b-5c51478a5d60', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '933dd0ba-3083-4fec-bc5f-1270e8761e92', 'vif_model': 'vmxnet3'}] {{(pid=63088) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 937.170344] env[63088]: DEBUG oslo.service.loopingcall [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 937.170666] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 79c4d6e8-9999-4b0c-98d3-bbfd4d61e660] Creating VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 937.170913] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-864a1193-6140-4d69-a2de-1e4d3876348d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.193820] env[63088]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 937.193820] env[63088]: value = "task-1285225" [ 937.193820] env[63088]: _type = "Task" [ 937.193820] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.203732] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285225, 'name': CreateVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.290769] env[63088]: DEBUG nova.compute.manager [req-7fe87168-93dd-4351-8514-3e608c5faab9 req-8af7101e-c6ba-4b50-a6e0-beea9f6ec6cd service nova] [instance: 79c4d6e8-9999-4b0c-98d3-bbfd4d61e660] Received event network-changed-933dd0ba-3083-4fec-bc5f-1270e8761e92 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 937.291444] env[63088]: DEBUG nova.compute.manager [req-7fe87168-93dd-4351-8514-3e608c5faab9 req-8af7101e-c6ba-4b50-a6e0-beea9f6ec6cd service nova] [instance: 79c4d6e8-9999-4b0c-98d3-bbfd4d61e660] Refreshing instance network info cache due to event network-changed-933dd0ba-3083-4fec-bc5f-1270e8761e92. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 937.298031] env[63088]: DEBUG oslo_concurrency.lockutils [req-7fe87168-93dd-4351-8514-3e608c5faab9 req-8af7101e-c6ba-4b50-a6e0-beea9f6ec6cd service nova] Acquiring lock "refresh_cache-79c4d6e8-9999-4b0c-98d3-bbfd4d61e660" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 937.298031] env[63088]: DEBUG oslo_concurrency.lockutils [req-7fe87168-93dd-4351-8514-3e608c5faab9 req-8af7101e-c6ba-4b50-a6e0-beea9f6ec6cd service nova] Acquired lock "refresh_cache-79c4d6e8-9999-4b0c-98d3-bbfd4d61e660" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 937.298031] env[63088]: DEBUG nova.network.neutron [req-7fe87168-93dd-4351-8514-3e608c5faab9 req-8af7101e-c6ba-4b50-a6e0-beea9f6ec6cd service nova] [instance: 79c4d6e8-9999-4b0c-98d3-bbfd4d61e660] Refreshing network info cache for port 933dd0ba-3083-4fec-bc5f-1270e8761e92 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 937.401917] env[63088]: DEBUG oslo_concurrency.lockutils [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.340s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.402617] env[63088]: DEBUG nova.compute.manager [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 7cd45e52-c6be-4ec6-9710-fa99e29cf872] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 937.408935] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5c7fe3f5-a517-4b12-9489-f6aa83255a48 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.130s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.408935] env[63088]: DEBUG nova.objects.instance [None req-5c7fe3f5-a517-4b12-9489-f6aa83255a48 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Lazy-loading 'resources' on Instance uuid dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 937.524077] env[63088]: DEBUG nova.compute.manager [req-26d1016b-9514-485e-82a2-f1d8b2de000b req-a8344661-e4a4-4a30-bfb4-14ee4e73f905 service nova] [instance: 2ad33254-8030-4454-b023-3359071f85de] Received event network-vif-plugged-1d2bee7e-bbc7-4b17-976b-47e7bbb268af {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 937.524351] env[63088]: DEBUG oslo_concurrency.lockutils [req-26d1016b-9514-485e-82a2-f1d8b2de000b req-a8344661-e4a4-4a30-bfb4-14ee4e73f905 service nova] Acquiring lock "2ad33254-8030-4454-b023-3359071f85de-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.524605] env[63088]: DEBUG oslo_concurrency.lockutils [req-26d1016b-9514-485e-82a2-f1d8b2de000b req-a8344661-e4a4-4a30-bfb4-14ee4e73f905 service nova] Lock "2ad33254-8030-4454-b023-3359071f85de-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.524873] env[63088]: DEBUG oslo_concurrency.lockutils [req-26d1016b-9514-485e-82a2-f1d8b2de000b req-a8344661-e4a4-4a30-bfb4-14ee4e73f905 service nova] Lock "2ad33254-8030-4454-b023-3359071f85de-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.525149] env[63088]: DEBUG nova.compute.manager [req-26d1016b-9514-485e-82a2-f1d8b2de000b req-a8344661-e4a4-4a30-bfb4-14ee4e73f905 service nova] [instance: 2ad33254-8030-4454-b023-3359071f85de] No waiting events found dispatching network-vif-plugged-1d2bee7e-bbc7-4b17-976b-47e7bbb268af {{(pid=63088) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 937.525338] env[63088]: WARNING nova.compute.manager [req-26d1016b-9514-485e-82a2-f1d8b2de000b req-a8344661-e4a4-4a30-bfb4-14ee4e73f905 service nova] [instance: 2ad33254-8030-4454-b023-3359071f85de] Received unexpected event network-vif-plugged-1d2bee7e-bbc7-4b17-976b-47e7bbb268af for instance with vm_state building and task_state spawning. [ 937.620510] env[63088]: DEBUG nova.network.neutron [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2ad33254-8030-4454-b023-3359071f85de] Successfully updated port: 1d2bee7e-bbc7-4b17-976b-47e7bbb268af {{(pid=63088) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 937.705384] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285225, 'name': CreateVM_Task, 'duration_secs': 0.328364} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.705584] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 79c4d6e8-9999-4b0c-98d3-bbfd4d61e660] Created VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 937.706406] env[63088]: DEBUG oslo_concurrency.lockutils [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c89225f6-2d34-4988-ae47-d3d6903890da" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 937.706882] env[63088]: DEBUG oslo_concurrency.lockutils [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c89225f6-2d34-4988-ae47-d3d6903890da" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 937.706882] env[63088]: DEBUG oslo_concurrency.lockutils [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c89225f6-2d34-4988-ae47-d3d6903890da" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 937.707200] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-69f41c33-d984-4969-80e1-dae75f3281a4 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.712273] env[63088]: DEBUG oslo_vmware.api [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Waiting for the task: (returnval){ [ 937.712273] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52ac9f0b-cea1-d942-84d8-ec8970abc9ae" [ 937.712273] env[63088]: _type = "Task" [ 937.712273] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.721486] env[63088]: DEBUG oslo_concurrency.lockutils [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Acquiring lock "49ee4180-6322-4003-a5b1-f2a91d190290" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.721736] env[63088]: DEBUG oslo_concurrency.lockutils [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Lock "49ee4180-6322-4003-a5b1-f2a91d190290" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.726565] env[63088]: DEBUG oslo_vmware.api [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52ac9f0b-cea1-d942-84d8-ec8970abc9ae, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.910684] env[63088]: DEBUG nova.compute.utils [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 937.917009] env[63088]: DEBUG nova.compute.manager [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 7cd45e52-c6be-4ec6-9710-fa99e29cf872] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 937.917209] env[63088]: DEBUG nova.network.neutron [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 7cd45e52-c6be-4ec6-9710-fa99e29cf872] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 938.039702] env[63088]: DEBUG nova.policy [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '32d5ff9b3ea345e4ac56010aa917e51b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '03e20d606e654362acbe2b36fe499ae3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 938.066645] env[63088]: DEBUG nova.network.neutron [req-7fe87168-93dd-4351-8514-3e608c5faab9 req-8af7101e-c6ba-4b50-a6e0-beea9f6ec6cd service nova] [instance: 79c4d6e8-9999-4b0c-98d3-bbfd4d61e660] Updated VIF entry in instance network info cache for port 933dd0ba-3083-4fec-bc5f-1270e8761e92. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 938.067142] env[63088]: DEBUG nova.network.neutron [req-7fe87168-93dd-4351-8514-3e608c5faab9 req-8af7101e-c6ba-4b50-a6e0-beea9f6ec6cd service nova] [instance: 79c4d6e8-9999-4b0c-98d3-bbfd4d61e660] Updating instance_info_cache with network_info: [{"id": "933dd0ba-3083-4fec-bc5f-1270e8761e92", "address": "fa:16:3e:4f:7c:f1", "network": {"id": "2bbaf351-fe63-46ba-ba7e-0f432c0a1bfb", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1438884226-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7bff1382b9694df08133c88a5fe783a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cca1f087-01e1-49ca-831b-5c51478a5d60", "external-id": "nsx-vlan-transportzone-439", "segmentation_id": 439, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap933dd0ba-30", "ovs_interfaceid": "933dd0ba-3083-4fec-bc5f-1270e8761e92", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 938.126816] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquiring lock "refresh_cache-2ad33254-8030-4454-b023-3359071f85de" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 938.127238] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquired lock "refresh_cache-2ad33254-8030-4454-b023-3359071f85de" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 938.128428] env[63088]: DEBUG nova.network.neutron [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2ad33254-8030-4454-b023-3359071f85de] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 938.149136] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7874ab69-2ae5-44e6-a673-439b2068488f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.157895] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f675aae-76a8-4f6c-9576-032be585278a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.191667] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cbe4140-9472-4f71-92f5-fd407c61d4a9 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.202337] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48a43291-9853-4741-a5ab-75417d293114 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.216744] env[63088]: DEBUG nova.compute.provider_tree [None req-5c7fe3f5-a517-4b12-9489-f6aa83255a48 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 938.227352] env[63088]: DEBUG oslo_concurrency.lockutils [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c89225f6-2d34-4988-ae47-d3d6903890da" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 938.227502] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 79c4d6e8-9999-4b0c-98d3-bbfd4d61e660] Processing image c89225f6-2d34-4988-ae47-d3d6903890da {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 938.227889] env[63088]: DEBUG oslo_concurrency.lockutils [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c89225f6-2d34-4988-ae47-d3d6903890da/c89225f6-2d34-4988-ae47-d3d6903890da.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 938.227889] env[63088]: DEBUG oslo_concurrency.lockutils [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c89225f6-2d34-4988-ae47-d3d6903890da/c89225f6-2d34-4988-ae47-d3d6903890da.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 938.228178] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 938.229092] env[63088]: DEBUG nova.compute.manager [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] [instance: 49ee4180-6322-4003-a5b1-f2a91d190290] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 938.231614] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-778b92a0-9119-4f92-aa9f-75ce368d36ae {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.242634] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 938.242634] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 938.243199] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-85a47843-3b18-498c-ba1a-b6e5f0ec4b8f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.249307] env[63088]: DEBUG oslo_vmware.api [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Waiting for the task: (returnval){ [ 938.249307] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]521c1d4d-8d13-0063-0777-c0544db37790" [ 938.249307] env[63088]: _type = "Task" [ 938.249307] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.257391] env[63088]: DEBUG oslo_vmware.api [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]521c1d4d-8d13-0063-0777-c0544db37790, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.418299] env[63088]: DEBUG nova.compute.manager [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 7cd45e52-c6be-4ec6-9710-fa99e29cf872] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 938.570717] env[63088]: DEBUG oslo_concurrency.lockutils [req-7fe87168-93dd-4351-8514-3e608c5faab9 req-8af7101e-c6ba-4b50-a6e0-beea9f6ec6cd service nova] Releasing lock "refresh_cache-79c4d6e8-9999-4b0c-98d3-bbfd4d61e660" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 938.597262] env[63088]: DEBUG nova.network.neutron [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 7cd45e52-c6be-4ec6-9710-fa99e29cf872] Successfully created port: ba2a84ec-889d-4dc1-b6ca-95d59271b186 {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 938.661197] env[63088]: DEBUG nova.network.neutron [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2ad33254-8030-4454-b023-3359071f85de] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 938.724321] env[63088]: DEBUG nova.scheduler.client.report [None req-5c7fe3f5-a517-4b12-9489-f6aa83255a48 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 938.755996] env[63088]: DEBUG oslo_concurrency.lockutils [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 938.760255] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 79c4d6e8-9999-4b0c-98d3-bbfd4d61e660] Preparing fetch location {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 938.760562] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 79c4d6e8-9999-4b0c-98d3-bbfd4d61e660] Fetch image to [datastore2] OSTACK_IMG_45886bbe-7f4c-4eff-8288-7270bbf01c81/OSTACK_IMG_45886bbe-7f4c-4eff-8288-7270bbf01c81.vmdk {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 938.760803] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 79c4d6e8-9999-4b0c-98d3-bbfd4d61e660] Downloading stream optimized image c89225f6-2d34-4988-ae47-d3d6903890da to [datastore2] OSTACK_IMG_45886bbe-7f4c-4eff-8288-7270bbf01c81/OSTACK_IMG_45886bbe-7f4c-4eff-8288-7270bbf01c81.vmdk on the data store datastore2 as vApp {{(pid=63088) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 938.761023] env[63088]: DEBUG nova.virt.vmwareapi.images [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 79c4d6e8-9999-4b0c-98d3-bbfd4d61e660] Downloading image file data c89225f6-2d34-4988-ae47-d3d6903890da to the ESX as VM named 'OSTACK_IMG_45886bbe-7f4c-4eff-8288-7270bbf01c81' {{(pid=63088) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 938.814685] env[63088]: DEBUG nova.network.neutron [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2ad33254-8030-4454-b023-3359071f85de] Updating instance_info_cache with network_info: [{"id": "1d2bee7e-bbc7-4b17-976b-47e7bbb268af", "address": "fa:16:3e:34:9a:ff", "network": {"id": "dff14a7f-0af1-4e4e-a498-86d7c9816e6b", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-762073671-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f33f2701fad94864a8c406a404bc0a42", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "21310d90-efbc-45a8-a97f-c4358606530f", "external-id": "nsx-vlan-transportzone-672", "segmentation_id": 672, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d2bee7e-bb", "ovs_interfaceid": "1d2bee7e-bbc7-4b17-976b-47e7bbb268af", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 938.843678] env[63088]: DEBUG oslo_vmware.rw_handles [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 938.843678] env[63088]: value = "resgroup-9" [ 938.843678] env[63088]: _type = "ResourcePool" [ 938.843678] env[63088]: }. {{(pid=63088) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 938.843980] env[63088]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-c873c7e1-453e-420f-bb8c-0a7633fc643c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.868740] env[63088]: DEBUG oslo_vmware.rw_handles [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Lease: (returnval){ [ 938.868740] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52336de1-9336-bc10-ca45-8d0b2abf1699" [ 938.868740] env[63088]: _type = "HttpNfcLease" [ 938.868740] env[63088]: } obtained for vApp import into resource pool (val){ [ 938.868740] env[63088]: value = "resgroup-9" [ 938.868740] env[63088]: _type = "ResourcePool" [ 938.868740] env[63088]: }. {{(pid=63088) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 938.869243] env[63088]: DEBUG oslo_vmware.api [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Waiting for the lease: (returnval){ [ 938.869243] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52336de1-9336-bc10-ca45-8d0b2abf1699" [ 938.869243] env[63088]: _type = "HttpNfcLease" [ 938.869243] env[63088]: } to be ready. {{(pid=63088) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 938.877435] env[63088]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 938.877435] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52336de1-9336-bc10-ca45-8d0b2abf1699" [ 938.877435] env[63088]: _type = "HttpNfcLease" [ 938.877435] env[63088]: } is initializing. {{(pid=63088) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 939.240558] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5c7fe3f5-a517-4b12-9489-f6aa83255a48 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.835s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 939.243122] env[63088]: DEBUG oslo_concurrency.lockutils [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.048s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 939.245201] env[63088]: INFO nova.compute.claims [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] [instance: 8205f9e1-effd-442d-990b-bf5a2a9e6cdd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 939.267992] env[63088]: INFO nova.scheduler.client.report [None req-5c7fe3f5-a517-4b12-9489-f6aa83255a48 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Deleted allocations for instance dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69 [ 939.317242] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Releasing lock "refresh_cache-2ad33254-8030-4454-b023-3359071f85de" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 939.317594] env[63088]: DEBUG nova.compute.manager [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2ad33254-8030-4454-b023-3359071f85de] Instance network_info: |[{"id": "1d2bee7e-bbc7-4b17-976b-47e7bbb268af", "address": "fa:16:3e:34:9a:ff", "network": {"id": "dff14a7f-0af1-4e4e-a498-86d7c9816e6b", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-762073671-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f33f2701fad94864a8c406a404bc0a42", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "21310d90-efbc-45a8-a97f-c4358606530f", "external-id": "nsx-vlan-transportzone-672", "segmentation_id": 672, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d2bee7e-bb", "ovs_interfaceid": "1d2bee7e-bbc7-4b17-976b-47e7bbb268af", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 939.318390] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2ad33254-8030-4454-b023-3359071f85de] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:34:9a:ff', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '21310d90-efbc-45a8-a97f-c4358606530f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1d2bee7e-bbc7-4b17-976b-47e7bbb268af', 'vif_model': 'vmxnet3'}] {{(pid=63088) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 939.326671] env[63088]: DEBUG oslo.service.loopingcall [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 939.327295] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2ad33254-8030-4454-b023-3359071f85de] Creating VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 939.327585] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7b4ba326-06f4-4108-9486-15963cc4523d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.349176] env[63088]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 939.349176] env[63088]: value = "task-1285227" [ 939.349176] env[63088]: _type = "Task" [ 939.349176] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.357185] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285227, 'name': CreateVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.376391] env[63088]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 939.376391] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52336de1-9336-bc10-ca45-8d0b2abf1699" [ 939.376391] env[63088]: _type = "HttpNfcLease" [ 939.376391] env[63088]: } is ready. {{(pid=63088) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 939.376717] env[63088]: DEBUG oslo_vmware.rw_handles [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 939.376717] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52336de1-9336-bc10-ca45-8d0b2abf1699" [ 939.376717] env[63088]: _type = "HttpNfcLease" [ 939.376717] env[63088]: }. {{(pid=63088) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 939.377457] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-daa4d6b8-1b2e-40fe-8de0-1387304c691b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.384688] env[63088]: DEBUG oslo_vmware.rw_handles [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52250d18-7b57-55b0-0ee3-572e439b3132/disk-0.vmdk from lease info. {{(pid=63088) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 939.384877] env[63088]: DEBUG oslo_vmware.rw_handles [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Creating HTTP connection to write to file with size = 21334016 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52250d18-7b57-55b0-0ee3-572e439b3132/disk-0.vmdk. {{(pid=63088) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 939.452030] env[63088]: DEBUG nova.compute.manager [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 7cd45e52-c6be-4ec6-9710-fa99e29cf872] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 939.458658] env[63088]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-f0c6fefa-4669-4db4-88f0-f5ec17e399df {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.650741] env[63088]: DEBUG nova.compute.manager [req-084f4d7c-0c33-4ce7-86c8-0e443088ad9e req-7707fee5-9757-48d4-a65f-b594c4f7134f service nova] [instance: 2ad33254-8030-4454-b023-3359071f85de] Received event network-changed-1d2bee7e-bbc7-4b17-976b-47e7bbb268af {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 939.651107] env[63088]: DEBUG nova.compute.manager [req-084f4d7c-0c33-4ce7-86c8-0e443088ad9e req-7707fee5-9757-48d4-a65f-b594c4f7134f service nova] [instance: 2ad33254-8030-4454-b023-3359071f85de] Refreshing instance network info cache due to event network-changed-1d2bee7e-bbc7-4b17-976b-47e7bbb268af. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 939.651420] env[63088]: DEBUG oslo_concurrency.lockutils [req-084f4d7c-0c33-4ce7-86c8-0e443088ad9e req-7707fee5-9757-48d4-a65f-b594c4f7134f service nova] Acquiring lock "refresh_cache-2ad33254-8030-4454-b023-3359071f85de" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 939.651609] env[63088]: DEBUG oslo_concurrency.lockutils [req-084f4d7c-0c33-4ce7-86c8-0e443088ad9e req-7707fee5-9757-48d4-a65f-b594c4f7134f service nova] Acquired lock "refresh_cache-2ad33254-8030-4454-b023-3359071f85de" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 939.651835] env[63088]: DEBUG nova.network.neutron [req-084f4d7c-0c33-4ce7-86c8-0e443088ad9e req-7707fee5-9757-48d4-a65f-b594c4f7134f service nova] [instance: 2ad33254-8030-4454-b023-3359071f85de] Refreshing network info cache for port 1d2bee7e-bbc7-4b17-976b-47e7bbb268af {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 939.779561] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5c7fe3f5-a517-4b12-9489-f6aa83255a48 tempest-AttachVolumeTestJSON-375388015 tempest-AttachVolumeTestJSON-375388015-project-member] Lock "dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.968s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 939.867320] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285227, 'name': CreateVM_Task, 'duration_secs': 0.340768} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.872067] env[63088]: DEBUG nova.virt.hardware [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 939.872324] env[63088]: DEBUG nova.virt.hardware [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 939.872544] env[63088]: DEBUG nova.virt.hardware [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 939.872806] env[63088]: DEBUG nova.virt.hardware [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 939.872995] env[63088]: DEBUG nova.virt.hardware [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 939.873210] env[63088]: DEBUG nova.virt.hardware [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 939.873484] env[63088]: DEBUG nova.virt.hardware [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 939.873708] env[63088]: DEBUG nova.virt.hardware [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 939.873875] env[63088]: DEBUG nova.virt.hardware [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 939.874058] env[63088]: DEBUG nova.virt.hardware [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 939.874320] env[63088]: DEBUG nova.virt.hardware [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 939.874607] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2ad33254-8030-4454-b023-3359071f85de] Created VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 939.875480] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8842f34-9b43-4ee6-95c9-6822bfc6437e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.880759] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 939.881017] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 939.881396] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 939.881780] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f398ad0d-21aa-4a54-89a7-839861f9ab07 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.891912] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfe55a56-60a1-4745-9139-db90e9afdea3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.896773] env[63088]: DEBUG oslo_vmware.api [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Waiting for the task: (returnval){ [ 939.896773] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52051b11-737a-9a3c-6bb8-0c01e723d927" [ 939.896773] env[63088]: _type = "Task" [ 939.896773] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.920386] env[63088]: DEBUG oslo_vmware.api [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52051b11-737a-9a3c-6bb8-0c01e723d927, 'name': SearchDatastore_Task, 'duration_secs': 0.011723} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.922239] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 939.922466] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2ad33254-8030-4454-b023-3359071f85de] Processing image 1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 939.922708] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 939.922878] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 939.923100] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 939.923435] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-64abcb25-d2ec-4242-b93a-260f925e7827 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.933630] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 939.933887] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 939.937460] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d83dbfa7-30db-4de9-96d7-c14b4f086a30 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.944907] env[63088]: DEBUG oslo_vmware.api [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Waiting for the task: (returnval){ [ 939.944907] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]526c59fc-75f5-a444-73ea-fed706cd98b1" [ 939.944907] env[63088]: _type = "Task" [ 939.944907] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.955923] env[63088]: DEBUG oslo_vmware.api [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]526c59fc-75f5-a444-73ea-fed706cd98b1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.162680] env[63088]: DEBUG nova.compute.manager [req-319ba3f1-6eb7-4090-abc7-e7c6e5513205 req-7024e759-1dcf-4717-b7b8-248f421e860b service nova] [instance: 7cd45e52-c6be-4ec6-9710-fa99e29cf872] Received event network-vif-plugged-ba2a84ec-889d-4dc1-b6ca-95d59271b186 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 940.163128] env[63088]: DEBUG oslo_concurrency.lockutils [req-319ba3f1-6eb7-4090-abc7-e7c6e5513205 req-7024e759-1dcf-4717-b7b8-248f421e860b service nova] Acquiring lock "7cd45e52-c6be-4ec6-9710-fa99e29cf872-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 940.163128] env[63088]: DEBUG oslo_concurrency.lockutils [req-319ba3f1-6eb7-4090-abc7-e7c6e5513205 req-7024e759-1dcf-4717-b7b8-248f421e860b service nova] Lock "7cd45e52-c6be-4ec6-9710-fa99e29cf872-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.163261] env[63088]: DEBUG oslo_concurrency.lockutils [req-319ba3f1-6eb7-4090-abc7-e7c6e5513205 req-7024e759-1dcf-4717-b7b8-248f421e860b service nova] Lock "7cd45e52-c6be-4ec6-9710-fa99e29cf872-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.163439] env[63088]: DEBUG nova.compute.manager [req-319ba3f1-6eb7-4090-abc7-e7c6e5513205 req-7024e759-1dcf-4717-b7b8-248f421e860b service nova] [instance: 7cd45e52-c6be-4ec6-9710-fa99e29cf872] No waiting events found dispatching network-vif-plugged-ba2a84ec-889d-4dc1-b6ca-95d59271b186 {{(pid=63088) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 940.163616] env[63088]: WARNING nova.compute.manager [req-319ba3f1-6eb7-4090-abc7-e7c6e5513205 req-7024e759-1dcf-4717-b7b8-248f421e860b service nova] [instance: 7cd45e52-c6be-4ec6-9710-fa99e29cf872] Received unexpected event network-vif-plugged-ba2a84ec-889d-4dc1-b6ca-95d59271b186 for instance with vm_state building and task_state spawning. [ 940.322599] env[63088]: DEBUG nova.network.neutron [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 7cd45e52-c6be-4ec6-9710-fa99e29cf872] Successfully updated port: ba2a84ec-889d-4dc1-b6ca-95d59271b186 {{(pid=63088) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 940.335037] env[63088]: DEBUG oslo_vmware.rw_handles [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Completed reading data from the image iterator. {{(pid=63088) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 940.335166] env[63088]: DEBUG oslo_vmware.rw_handles [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52250d18-7b57-55b0-0ee3-572e439b3132/disk-0.vmdk. {{(pid=63088) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 940.336145] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96071231-d03c-4ba9-b28f-b14beff977d9 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.344304] env[63088]: DEBUG oslo_vmware.rw_handles [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52250d18-7b57-55b0-0ee3-572e439b3132/disk-0.vmdk is in state: ready. {{(pid=63088) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 940.344495] env[63088]: DEBUG oslo_vmware.rw_handles [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Releasing lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52250d18-7b57-55b0-0ee3-572e439b3132/disk-0.vmdk. {{(pid=63088) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 940.344755] env[63088]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-dc633218-899b-4d04-980e-74175ec9d742 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.459882] env[63088]: DEBUG oslo_vmware.api [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]526c59fc-75f5-a444-73ea-fed706cd98b1, 'name': SearchDatastore_Task, 'duration_secs': 0.011165} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.461462] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-144dbad8-bdda-4e40-a7f4-65e3ba20b22e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.469405] env[63088]: DEBUG oslo_vmware.api [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Waiting for the task: (returnval){ [ 940.469405] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]522479fb-199c-c5f4-6e5a-2758dae6a285" [ 940.469405] env[63088]: _type = "Task" [ 940.469405] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.478859] env[63088]: DEBUG oslo_vmware.api [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]522479fb-199c-c5f4-6e5a-2758dae6a285, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.479853] env[63088]: DEBUG nova.network.neutron [req-084f4d7c-0c33-4ce7-86c8-0e443088ad9e req-7707fee5-9757-48d4-a65f-b594c4f7134f service nova] [instance: 2ad33254-8030-4454-b023-3359071f85de] Updated VIF entry in instance network info cache for port 1d2bee7e-bbc7-4b17-976b-47e7bbb268af. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 940.480246] env[63088]: DEBUG nova.network.neutron [req-084f4d7c-0c33-4ce7-86c8-0e443088ad9e req-7707fee5-9757-48d4-a65f-b594c4f7134f service nova] [instance: 2ad33254-8030-4454-b023-3359071f85de] Updating instance_info_cache with network_info: [{"id": "1d2bee7e-bbc7-4b17-976b-47e7bbb268af", "address": "fa:16:3e:34:9a:ff", "network": {"id": "dff14a7f-0af1-4e4e-a498-86d7c9816e6b", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-762073671-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f33f2701fad94864a8c406a404bc0a42", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "21310d90-efbc-45a8-a97f-c4358606530f", "external-id": "nsx-vlan-transportzone-672", "segmentation_id": 672, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d2bee7e-bb", "ovs_interfaceid": "1d2bee7e-bbc7-4b17-976b-47e7bbb268af", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 940.484119] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67623afe-9335-497f-90a9-a3efa9e6ca0d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.490485] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71215b1f-a776-4069-b4ca-7c8baac11fb8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.525967] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-069c6c13-7c85-4c70-99cc-98734b60a487 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.535249] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1e052a6-bdc2-43e2-9fba-38e1d76d6ec5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.551546] env[63088]: DEBUG nova.compute.provider_tree [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 940.604677] env[63088]: DEBUG oslo_concurrency.lockutils [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Acquiring lock "4a56d74c-cb1f-4edf-99f0-4b54bf357a3a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 940.604924] env[63088]: DEBUG oslo_concurrency.lockutils [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Lock "4a56d74c-cb1f-4edf-99f0-4b54bf357a3a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.643751] env[63088]: DEBUG oslo_vmware.rw_handles [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Closed VMDK write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52250d18-7b57-55b0-0ee3-572e439b3132/disk-0.vmdk. {{(pid=63088) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 940.644044] env[63088]: INFO nova.virt.vmwareapi.images [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 79c4d6e8-9999-4b0c-98d3-bbfd4d61e660] Downloaded image file data c89225f6-2d34-4988-ae47-d3d6903890da [ 940.644900] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-836958df-b796-4484-95f6-6e1dc5e01680 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.661975] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4491e9ca-312d-49cf-abe5-57b54f3bdd90 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.689066] env[63088]: INFO nova.virt.vmwareapi.images [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 79c4d6e8-9999-4b0c-98d3-bbfd4d61e660] The imported VM was unregistered [ 940.691117] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 79c4d6e8-9999-4b0c-98d3-bbfd4d61e660] Caching image {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 940.691374] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Creating directory with path [datastore2] devstack-image-cache_base/c89225f6-2d34-4988-ae47-d3d6903890da {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 940.691686] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3af83aee-227d-4512-a410-6cc92ac2e07a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.703144] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Created directory with path [datastore2] devstack-image-cache_base/c89225f6-2d34-4988-ae47-d3d6903890da {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 940.703362] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_45886bbe-7f4c-4eff-8288-7270bbf01c81/OSTACK_IMG_45886bbe-7f4c-4eff-8288-7270bbf01c81.vmdk to [datastore2] devstack-image-cache_base/c89225f6-2d34-4988-ae47-d3d6903890da/c89225f6-2d34-4988-ae47-d3d6903890da.vmdk. {{(pid=63088) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 940.703655] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-841faa89-fa65-4148-881f-5b3586ba0487 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.710873] env[63088]: DEBUG oslo_vmware.api [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Waiting for the task: (returnval){ [ 940.710873] env[63088]: value = "task-1285230" [ 940.710873] env[63088]: _type = "Task" [ 940.710873] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.719110] env[63088]: DEBUG oslo_vmware.api [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285230, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.834440] env[63088]: DEBUG oslo_concurrency.lockutils [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquiring lock "refresh_cache-7cd45e52-c6be-4ec6-9710-fa99e29cf872" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 940.834616] env[63088]: DEBUG oslo_concurrency.lockutils [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquired lock "refresh_cache-7cd45e52-c6be-4ec6-9710-fa99e29cf872" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 940.834758] env[63088]: DEBUG nova.network.neutron [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 7cd45e52-c6be-4ec6-9710-fa99e29cf872] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 940.983086] env[63088]: DEBUG oslo_vmware.api [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]522479fb-199c-c5f4-6e5a-2758dae6a285, 'name': SearchDatastore_Task, 'duration_secs': 0.011101} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.983438] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 940.983742] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore2] 2ad33254-8030-4454-b023-3359071f85de/2ad33254-8030-4454-b023-3359071f85de.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 940.984096] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b3a80f2d-8df3-46ae-a54a-a04c0fb60c97 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.987136] env[63088]: DEBUG oslo_concurrency.lockutils [req-084f4d7c-0c33-4ce7-86c8-0e443088ad9e req-7707fee5-9757-48d4-a65f-b594c4f7134f service nova] Releasing lock "refresh_cache-2ad33254-8030-4454-b023-3359071f85de" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 940.994934] env[63088]: DEBUG oslo_vmware.api [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Waiting for the task: (returnval){ [ 940.994934] env[63088]: value = "task-1285231" [ 940.994934] env[63088]: _type = "Task" [ 940.994934] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.005291] env[63088]: DEBUG oslo_vmware.api [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285231, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.054624] env[63088]: DEBUG nova.scheduler.client.report [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 941.109050] env[63088]: DEBUG nova.compute.manager [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] [instance: 4a56d74c-cb1f-4edf-99f0-4b54bf357a3a] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 941.224084] env[63088]: DEBUG oslo_vmware.api [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285230, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.368402] env[63088]: DEBUG nova.network.neutron [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 7cd45e52-c6be-4ec6-9710-fa99e29cf872] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 941.506602] env[63088]: DEBUG oslo_vmware.api [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285231, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.561286] env[63088]: DEBUG oslo_concurrency.lockutils [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.318s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.561666] env[63088]: DEBUG nova.compute.manager [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] [instance: 8205f9e1-effd-442d-990b-bf5a2a9e6cdd] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 941.566641] env[63088]: DEBUG oslo_concurrency.lockutils [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.741s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 941.569901] env[63088]: INFO nova.compute.claims [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 941.636982] env[63088]: DEBUG oslo_concurrency.lockutils [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.724014] env[63088]: DEBUG oslo_vmware.api [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285230, 'name': MoveVirtualDisk_Task} progress is 43%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.808145] env[63088]: DEBUG nova.network.neutron [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 7cd45e52-c6be-4ec6-9710-fa99e29cf872] Updating instance_info_cache with network_info: [{"id": "ba2a84ec-889d-4dc1-b6ca-95d59271b186", "address": "fa:16:3e:f1:9d:80", "network": {"id": "307b966a-d9e1-40fd-9313-1ad94c734308", "bridge": "br-int", "label": "tempest-ServersTestJSON-1907249197-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "03e20d606e654362acbe2b36fe499ae3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae18b41f-e73c-44f1-83dd-467c080944f4", "external-id": "nsx-vlan-transportzone-653", "segmentation_id": 653, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapba2a84ec-88", "ovs_interfaceid": "ba2a84ec-889d-4dc1-b6ca-95d59271b186", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 942.009879] env[63088]: DEBUG oslo_vmware.api [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285231, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.075926] env[63088]: DEBUG nova.compute.utils [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 942.078554] env[63088]: DEBUG nova.compute.manager [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] [instance: 8205f9e1-effd-442d-990b-bf5a2a9e6cdd] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 942.078554] env[63088]: DEBUG nova.network.neutron [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] [instance: 8205f9e1-effd-442d-990b-bf5a2a9e6cdd] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 942.132048] env[63088]: DEBUG nova.policy [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f002e9ea069a4f348e65f0578f0f6df1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f95a6d07d1ba4f618b3c6bd3960353fd', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 942.191617] env[63088]: DEBUG nova.compute.manager [req-c802cd45-c8c0-4016-9ff7-82785080b664 req-9acea611-8073-4056-a3c6-c1a386d0ede2 service nova] [instance: 7cd45e52-c6be-4ec6-9710-fa99e29cf872] Received event network-changed-ba2a84ec-889d-4dc1-b6ca-95d59271b186 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 942.191873] env[63088]: DEBUG nova.compute.manager [req-c802cd45-c8c0-4016-9ff7-82785080b664 req-9acea611-8073-4056-a3c6-c1a386d0ede2 service nova] [instance: 7cd45e52-c6be-4ec6-9710-fa99e29cf872] Refreshing instance network info cache due to event network-changed-ba2a84ec-889d-4dc1-b6ca-95d59271b186. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 942.192262] env[63088]: DEBUG oslo_concurrency.lockutils [req-c802cd45-c8c0-4016-9ff7-82785080b664 req-9acea611-8073-4056-a3c6-c1a386d0ede2 service nova] Acquiring lock "refresh_cache-7cd45e52-c6be-4ec6-9710-fa99e29cf872" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 942.228626] env[63088]: DEBUG oslo_vmware.api [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285230, 'name': MoveVirtualDisk_Task} progress is 66%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.311708] env[63088]: DEBUG oslo_concurrency.lockutils [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Releasing lock "refresh_cache-7cd45e52-c6be-4ec6-9710-fa99e29cf872" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 942.312043] env[63088]: DEBUG nova.compute.manager [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 7cd45e52-c6be-4ec6-9710-fa99e29cf872] Instance network_info: |[{"id": "ba2a84ec-889d-4dc1-b6ca-95d59271b186", "address": "fa:16:3e:f1:9d:80", "network": {"id": "307b966a-d9e1-40fd-9313-1ad94c734308", "bridge": "br-int", "label": "tempest-ServersTestJSON-1907249197-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "03e20d606e654362acbe2b36fe499ae3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae18b41f-e73c-44f1-83dd-467c080944f4", "external-id": "nsx-vlan-transportzone-653", "segmentation_id": 653, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapba2a84ec-88", "ovs_interfaceid": "ba2a84ec-889d-4dc1-b6ca-95d59271b186", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 942.312395] env[63088]: DEBUG oslo_concurrency.lockutils [req-c802cd45-c8c0-4016-9ff7-82785080b664 req-9acea611-8073-4056-a3c6-c1a386d0ede2 service nova] Acquired lock "refresh_cache-7cd45e52-c6be-4ec6-9710-fa99e29cf872" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 942.312681] env[63088]: DEBUG nova.network.neutron [req-c802cd45-c8c0-4016-9ff7-82785080b664 req-9acea611-8073-4056-a3c6-c1a386d0ede2 service nova] [instance: 7cd45e52-c6be-4ec6-9710-fa99e29cf872] Refreshing network info cache for port ba2a84ec-889d-4dc1-b6ca-95d59271b186 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 942.314802] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 7cd45e52-c6be-4ec6-9710-fa99e29cf872] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f1:9d:80', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ae18b41f-e73c-44f1-83dd-467c080944f4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ba2a84ec-889d-4dc1-b6ca-95d59271b186', 'vif_model': 'vmxnet3'}] {{(pid=63088) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 942.323776] env[63088]: DEBUG oslo.service.loopingcall [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 942.325050] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7cd45e52-c6be-4ec6-9710-fa99e29cf872] Creating VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 942.325726] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-68f3cf83-4090-42b9-bf1e-7cb5eec50a3f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.352143] env[63088]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 942.352143] env[63088]: value = "task-1285232" [ 942.352143] env[63088]: _type = "Task" [ 942.352143] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.364989] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285232, 'name': CreateVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.420477] env[63088]: DEBUG nova.network.neutron [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] [instance: 8205f9e1-effd-442d-990b-bf5a2a9e6cdd] Successfully created port: faac8e4c-2334-4c13-ae85-13e55b44fbb6 {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 942.509748] env[63088]: DEBUG oslo_vmware.api [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285231, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.582241] env[63088]: DEBUG nova.compute.manager [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] [instance: 8205f9e1-effd-442d-990b-bf5a2a9e6cdd] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 942.733158] env[63088]: DEBUG oslo_vmware.api [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285230, 'name': MoveVirtualDisk_Task} progress is 85%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.872589] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285232, 'name': CreateVM_Task} progress is 25%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.884660] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4aa03c43-1348-4f86-8f51-3b567cd4bcc2 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.896449] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b402e2d-656a-4b3d-80af-8545d9fd8d1a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.937827] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c273e7bb-0827-487b-aa59-5f885ff37eb4 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.948695] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a2fd8e6-4e81-4963-b456-4e516aa04616 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.972546] env[63088]: DEBUG nova.compute.provider_tree [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 943.010851] env[63088]: DEBUG oslo_vmware.api [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285231, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.210755] env[63088]: DEBUG nova.network.neutron [req-c802cd45-c8c0-4016-9ff7-82785080b664 req-9acea611-8073-4056-a3c6-c1a386d0ede2 service nova] [instance: 7cd45e52-c6be-4ec6-9710-fa99e29cf872] Updated VIF entry in instance network info cache for port ba2a84ec-889d-4dc1-b6ca-95d59271b186. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 943.211676] env[63088]: DEBUG nova.network.neutron [req-c802cd45-c8c0-4016-9ff7-82785080b664 req-9acea611-8073-4056-a3c6-c1a386d0ede2 service nova] [instance: 7cd45e52-c6be-4ec6-9710-fa99e29cf872] Updating instance_info_cache with network_info: [{"id": "ba2a84ec-889d-4dc1-b6ca-95d59271b186", "address": "fa:16:3e:f1:9d:80", "network": {"id": "307b966a-d9e1-40fd-9313-1ad94c734308", "bridge": "br-int", "label": "tempest-ServersTestJSON-1907249197-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "03e20d606e654362acbe2b36fe499ae3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae18b41f-e73c-44f1-83dd-467c080944f4", "external-id": "nsx-vlan-transportzone-653", "segmentation_id": 653, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapba2a84ec-88", "ovs_interfaceid": "ba2a84ec-889d-4dc1-b6ca-95d59271b186", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 943.226877] env[63088]: DEBUG oslo_vmware.api [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285230, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.364930] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285232, 'name': CreateVM_Task, 'duration_secs': 0.795837} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.365145] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7cd45e52-c6be-4ec6-9710-fa99e29cf872] Created VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 943.365797] env[63088]: DEBUG oslo_concurrency.lockutils [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 943.365966] env[63088]: DEBUG oslo_concurrency.lockutils [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.366324] env[63088]: DEBUG oslo_concurrency.lockutils [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 943.366584] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9b2bb3a1-d037-4224-8a3d-83c270327970 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.371336] env[63088]: DEBUG oslo_vmware.api [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Waiting for the task: (returnval){ [ 943.371336] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]523e4644-1f06-e4b3-e9ab-275e808a611b" [ 943.371336] env[63088]: _type = "Task" [ 943.371336] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.380026] env[63088]: DEBUG oslo_vmware.api [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]523e4644-1f06-e4b3-e9ab-275e808a611b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.479336] env[63088]: DEBUG nova.scheduler.client.report [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 943.508525] env[63088]: DEBUG oslo_vmware.api [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285231, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.310173} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.508796] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore2] 2ad33254-8030-4454-b023-3359071f85de/2ad33254-8030-4454-b023-3359071f85de.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 943.509015] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2ad33254-8030-4454-b023-3359071f85de] Extending root virtual disk to 1048576 {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 943.509280] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-51ec2956-57b3-4044-9b14-042a4da07b41 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.515848] env[63088]: DEBUG oslo_vmware.api [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Waiting for the task: (returnval){ [ 943.515848] env[63088]: value = "task-1285234" [ 943.515848] env[63088]: _type = "Task" [ 943.515848] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.524247] env[63088]: DEBUG oslo_vmware.api [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285234, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.596721] env[63088]: DEBUG nova.compute.manager [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] [instance: 8205f9e1-effd-442d-990b-bf5a2a9e6cdd] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 943.622489] env[63088]: DEBUG nova.virt.hardware [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 943.622787] env[63088]: DEBUG nova.virt.hardware [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 943.622950] env[63088]: DEBUG nova.virt.hardware [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 943.623153] env[63088]: DEBUG nova.virt.hardware [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 943.623307] env[63088]: DEBUG nova.virt.hardware [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 943.623459] env[63088]: DEBUG nova.virt.hardware [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 943.623665] env[63088]: DEBUG nova.virt.hardware [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 943.623855] env[63088]: DEBUG nova.virt.hardware [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 943.624077] env[63088]: DEBUG nova.virt.hardware [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 943.624258] env[63088]: DEBUG nova.virt.hardware [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 943.624489] env[63088]: DEBUG nova.virt.hardware [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 943.625738] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5fd2ff3-5b5b-42ad-8e32-9917f6b78f92 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.634347] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e249ddd-815d-452e-8d05-b856848fb760 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.719435] env[63088]: DEBUG oslo_concurrency.lockutils [req-c802cd45-c8c0-4016-9ff7-82785080b664 req-9acea611-8073-4056-a3c6-c1a386d0ede2 service nova] Releasing lock "refresh_cache-7cd45e52-c6be-4ec6-9710-fa99e29cf872" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 943.723986] env[63088]: DEBUG oslo_vmware.api [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285230, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.619795} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.724262] env[63088]: INFO nova.virt.vmwareapi.ds_util [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_45886bbe-7f4c-4eff-8288-7270bbf01c81/OSTACK_IMG_45886bbe-7f4c-4eff-8288-7270bbf01c81.vmdk to [datastore2] devstack-image-cache_base/c89225f6-2d34-4988-ae47-d3d6903890da/c89225f6-2d34-4988-ae47-d3d6903890da.vmdk. [ 943.724518] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 79c4d6e8-9999-4b0c-98d3-bbfd4d61e660] Cleaning up location [datastore2] OSTACK_IMG_45886bbe-7f4c-4eff-8288-7270bbf01c81 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 943.724627] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_45886bbe-7f4c-4eff-8288-7270bbf01c81 {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 943.724897] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-42108308-a35c-45f0-ba9f-57933f0e8d27 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.731217] env[63088]: DEBUG oslo_vmware.api [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Waiting for the task: (returnval){ [ 943.731217] env[63088]: value = "task-1285235" [ 943.731217] env[63088]: _type = "Task" [ 943.731217] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.738839] env[63088]: DEBUG oslo_vmware.api [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285235, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.878925] env[63088]: DEBUG nova.compute.manager [req-b4815cb8-6fe9-4f9a-854a-851f2bb2ff63 req-88c1033a-87ec-4fd4-9936-b2206b8013b8 service nova] [instance: 8205f9e1-effd-442d-990b-bf5a2a9e6cdd] Received event network-vif-plugged-faac8e4c-2334-4c13-ae85-13e55b44fbb6 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 943.879329] env[63088]: DEBUG oslo_concurrency.lockutils [req-b4815cb8-6fe9-4f9a-854a-851f2bb2ff63 req-88c1033a-87ec-4fd4-9936-b2206b8013b8 service nova] Acquiring lock "8205f9e1-effd-442d-990b-bf5a2a9e6cdd-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 943.879668] env[63088]: DEBUG oslo_concurrency.lockutils [req-b4815cb8-6fe9-4f9a-854a-851f2bb2ff63 req-88c1033a-87ec-4fd4-9936-b2206b8013b8 service nova] Lock "8205f9e1-effd-442d-990b-bf5a2a9e6cdd-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 943.879963] env[63088]: DEBUG oslo_concurrency.lockutils [req-b4815cb8-6fe9-4f9a-854a-851f2bb2ff63 req-88c1033a-87ec-4fd4-9936-b2206b8013b8 service nova] Lock "8205f9e1-effd-442d-990b-bf5a2a9e6cdd-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 943.880626] env[63088]: DEBUG nova.compute.manager [req-b4815cb8-6fe9-4f9a-854a-851f2bb2ff63 req-88c1033a-87ec-4fd4-9936-b2206b8013b8 service nova] [instance: 8205f9e1-effd-442d-990b-bf5a2a9e6cdd] No waiting events found dispatching network-vif-plugged-faac8e4c-2334-4c13-ae85-13e55b44fbb6 {{(pid=63088) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 943.880907] env[63088]: WARNING nova.compute.manager [req-b4815cb8-6fe9-4f9a-854a-851f2bb2ff63 req-88c1033a-87ec-4fd4-9936-b2206b8013b8 service nova] [instance: 8205f9e1-effd-442d-990b-bf5a2a9e6cdd] Received unexpected event network-vif-plugged-faac8e4c-2334-4c13-ae85-13e55b44fbb6 for instance with vm_state building and task_state spawning. [ 943.885434] env[63088]: DEBUG oslo_vmware.api [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]523e4644-1f06-e4b3-e9ab-275e808a611b, 'name': SearchDatastore_Task, 'duration_secs': 0.012237} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.885858] env[63088]: DEBUG oslo_concurrency.lockutils [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 943.886214] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 7cd45e52-c6be-4ec6-9710-fa99e29cf872] Processing image 1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 943.886558] env[63088]: DEBUG oslo_concurrency.lockutils [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 943.886814] env[63088]: DEBUG oslo_concurrency.lockutils [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.888016] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 943.888016] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-810ee7f3-5b63-4dde-88fa-2af10b7e8f0c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.905556] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 943.905927] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 943.907057] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d0d3cb2f-98c6-4a10-a8c6-1e157d5e6316 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.913894] env[63088]: DEBUG oslo_vmware.api [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Waiting for the task: (returnval){ [ 943.913894] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52be8aa7-ecd7-b708-6678-396770f124d7" [ 943.913894] env[63088]: _type = "Task" [ 943.913894] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.922134] env[63088]: DEBUG oslo_vmware.api [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52be8aa7-ecd7-b708-6678-396770f124d7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.974125] env[63088]: DEBUG nova.network.neutron [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] [instance: 8205f9e1-effd-442d-990b-bf5a2a9e6cdd] Successfully updated port: faac8e4c-2334-4c13-ae85-13e55b44fbb6 {{(pid=63088) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 943.984854] env[63088]: DEBUG oslo_concurrency.lockutils [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.420s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 943.985406] env[63088]: DEBUG nova.compute.manager [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 943.988649] env[63088]: DEBUG oslo_concurrency.lockutils [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.233s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 943.990273] env[63088]: INFO nova.compute.claims [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] [instance: 49ee4180-6322-4003-a5b1-f2a91d190290] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 944.028376] env[63088]: DEBUG oslo_vmware.api [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285234, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.240737] env[63088]: DEBUG oslo_vmware.api [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285235, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.219049} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.241031] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 944.241311] env[63088]: DEBUG oslo_concurrency.lockutils [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c89225f6-2d34-4988-ae47-d3d6903890da/c89225f6-2d34-4988-ae47-d3d6903890da.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 944.241926] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c89225f6-2d34-4988-ae47-d3d6903890da/c89225f6-2d34-4988-ae47-d3d6903890da.vmdk to [datastore2] 79c4d6e8-9999-4b0c-98d3-bbfd4d61e660/79c4d6e8-9999-4b0c-98d3-bbfd4d61e660.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 944.241926] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-19f6bce3-e10c-4c75-bf88-2827e3e949cc {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.248933] env[63088]: DEBUG oslo_vmware.api [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Waiting for the task: (returnval){ [ 944.248933] env[63088]: value = "task-1285236" [ 944.248933] env[63088]: _type = "Task" [ 944.248933] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.256619] env[63088]: DEBUG oslo_vmware.api [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285236, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.424899] env[63088]: DEBUG oslo_vmware.api [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52be8aa7-ecd7-b708-6678-396770f124d7, 'name': SearchDatastore_Task, 'duration_secs': 0.034595} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.425802] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-69737e30-b37c-483b-878d-45bf4a746565 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.431360] env[63088]: DEBUG oslo_vmware.api [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Waiting for the task: (returnval){ [ 944.431360] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52a3e296-2f4d-4564-80b3-b5f0bbe052e5" [ 944.431360] env[63088]: _type = "Task" [ 944.431360] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.439426] env[63088]: DEBUG oslo_vmware.api [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52a3e296-2f4d-4564-80b3-b5f0bbe052e5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.476338] env[63088]: DEBUG oslo_concurrency.lockutils [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Acquiring lock "refresh_cache-8205f9e1-effd-442d-990b-bf5a2a9e6cdd" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 944.476590] env[63088]: DEBUG oslo_concurrency.lockutils [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Acquired lock "refresh_cache-8205f9e1-effd-442d-990b-bf5a2a9e6cdd" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 944.477517] env[63088]: DEBUG nova.network.neutron [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] [instance: 8205f9e1-effd-442d-990b-bf5a2a9e6cdd] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 944.494490] env[63088]: DEBUG nova.compute.utils [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 944.498544] env[63088]: DEBUG nova.compute.manager [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 944.498690] env[63088]: DEBUG nova.network.neutron [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 944.527651] env[63088]: DEBUG oslo_vmware.api [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285234, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.641079} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.527987] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2ad33254-8030-4454-b023-3359071f85de] Extended root virtual disk {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 944.528986] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a13ba4e3-8372-4032-ac23-418fb6273097 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.554296] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2ad33254-8030-4454-b023-3359071f85de] Reconfiguring VM instance instance-00000055 to attach disk [datastore2] 2ad33254-8030-4454-b023-3359071f85de/2ad33254-8030-4454-b023-3359071f85de.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 944.555981] env[63088]: DEBUG nova.policy [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '93f1d334e71a4a99ba300c3ee6d70243', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'aa3d24a1a6c0430985fd80365d986ee1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 944.557561] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-49094f1d-8bc7-4960-af7a-fb2dfccb296f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.578690] env[63088]: DEBUG oslo_vmware.api [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Waiting for the task: (returnval){ [ 944.578690] env[63088]: value = "task-1285237" [ 944.578690] env[63088]: _type = "Task" [ 944.578690] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.588389] env[63088]: DEBUG oslo_vmware.api [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285237, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.761496] env[63088]: DEBUG oslo_vmware.api [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285236, 'name': CopyVirtualDisk_Task} progress is 18%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.875318] env[63088]: DEBUG nova.network.neutron [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Successfully created port: b7a01048-9168-46ff-a5b8-83cee3dd3c97 {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 944.943312] env[63088]: DEBUG oslo_vmware.api [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52a3e296-2f4d-4564-80b3-b5f0bbe052e5, 'name': SearchDatastore_Task, 'duration_secs': 0.012059} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.943674] env[63088]: DEBUG oslo_concurrency.lockutils [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 944.943991] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore2] 7cd45e52-c6be-4ec6-9710-fa99e29cf872/7cd45e52-c6be-4ec6-9710-fa99e29cf872.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 944.944311] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-31665554-48bc-445e-ab16-0679911b54e9 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.956804] env[63088]: DEBUG oslo_vmware.api [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Waiting for the task: (returnval){ [ 944.956804] env[63088]: value = "task-1285238" [ 944.956804] env[63088]: _type = "Task" [ 944.956804] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.968661] env[63088]: DEBUG oslo_vmware.api [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285238, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.999447] env[63088]: DEBUG nova.compute.manager [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 945.048154] env[63088]: DEBUG nova.network.neutron [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] [instance: 8205f9e1-effd-442d-990b-bf5a2a9e6cdd] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 945.093825] env[63088]: DEBUG oslo_vmware.api [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285237, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.264841] env[63088]: DEBUG oslo_vmware.api [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285236, 'name': CopyVirtualDisk_Task} progress is 40%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.265633] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d1cf929-c8a2-4d36-9a7d-4e72caa4480d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.276522] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fc40f31-027a-4a02-9e9a-ba528b1c3951 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.317702] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbece208-922b-4aec-a38b-b17ae24f162b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.328190] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c399747-f138-4a2b-816e-6859875c0afc {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.345356] env[63088]: DEBUG nova.compute.provider_tree [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 945.348202] env[63088]: DEBUG nova.network.neutron [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] [instance: 8205f9e1-effd-442d-990b-bf5a2a9e6cdd] Updating instance_info_cache with network_info: [{"id": "faac8e4c-2334-4c13-ae85-13e55b44fbb6", "address": "fa:16:3e:5f:f6:fa", "network": {"id": "cca88867-6df4-49c9-baec-a37aa8156a46", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-1660940561-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f95a6d07d1ba4f618b3c6bd3960353fd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4813d311-2016-4f6a-a4b8-7613ab624fba", "external-id": "nsx-vlan-transportzone-386", "segmentation_id": 386, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfaac8e4c-23", "ovs_interfaceid": "faac8e4c-2334-4c13-ae85-13e55b44fbb6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 945.470497] env[63088]: DEBUG oslo_vmware.rw_handles [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526bdc2d-92d1-17cb-9ddc-6c73bb412339/disk-0.vmdk. {{(pid=63088) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 945.471547] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44a15f82-f0d3-4602-b995-521574fbaf60 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.478067] env[63088]: DEBUG oslo_vmware.api [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285238, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.482337] env[63088]: DEBUG oslo_vmware.rw_handles [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526bdc2d-92d1-17cb-9ddc-6c73bb412339/disk-0.vmdk is in state: ready. {{(pid=63088) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 945.482585] env[63088]: ERROR oslo_vmware.rw_handles [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526bdc2d-92d1-17cb-9ddc-6c73bb412339/disk-0.vmdk due to incomplete transfer. [ 945.482864] env[63088]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-6f6df3f6-81b6-4b03-82a7-0e743675d91b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.490465] env[63088]: DEBUG oslo_vmware.rw_handles [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526bdc2d-92d1-17cb-9ddc-6c73bb412339/disk-0.vmdk. {{(pid=63088) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 945.490693] env[63088]: DEBUG nova.virt.vmwareapi.images [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] Uploaded image e0b4e03a-baf3-4fa2-ba4f-555963d990d6 to the Glance image server {{(pid=63088) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 945.493089] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] Destroying the VM {{(pid=63088) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 945.493409] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-03c10b00-75c7-416b-96bf-93a85dd4efa5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.500778] env[63088]: DEBUG oslo_vmware.api [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 945.500778] env[63088]: value = "task-1285239" [ 945.500778] env[63088]: _type = "Task" [ 945.500778] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.513084] env[63088]: DEBUG oslo_vmware.api [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285239, 'name': Destroy_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.592380] env[63088]: DEBUG oslo_vmware.api [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285237, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.760891] env[63088]: DEBUG oslo_vmware.api [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285236, 'name': CopyVirtualDisk_Task} progress is 60%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.849036] env[63088]: DEBUG nova.scheduler.client.report [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 945.854540] env[63088]: DEBUG oslo_concurrency.lockutils [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Releasing lock "refresh_cache-8205f9e1-effd-442d-990b-bf5a2a9e6cdd" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 945.854957] env[63088]: DEBUG nova.compute.manager [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] [instance: 8205f9e1-effd-442d-990b-bf5a2a9e6cdd] Instance network_info: |[{"id": "faac8e4c-2334-4c13-ae85-13e55b44fbb6", "address": "fa:16:3e:5f:f6:fa", "network": {"id": "cca88867-6df4-49c9-baec-a37aa8156a46", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-1660940561-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f95a6d07d1ba4f618b3c6bd3960353fd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4813d311-2016-4f6a-a4b8-7613ab624fba", "external-id": "nsx-vlan-transportzone-386", "segmentation_id": 386, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfaac8e4c-23", "ovs_interfaceid": "faac8e4c-2334-4c13-ae85-13e55b44fbb6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 945.855454] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] [instance: 8205f9e1-effd-442d-990b-bf5a2a9e6cdd] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5f:f6:fa', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4813d311-2016-4f6a-a4b8-7613ab624fba', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'faac8e4c-2334-4c13-ae85-13e55b44fbb6', 'vif_model': 'vmxnet3'}] {{(pid=63088) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 945.866357] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Creating folder: Project (f95a6d07d1ba4f618b3c6bd3960353fd). Parent ref: group-v275816. {{(pid=63088) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 945.868029] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1775b32b-76e1-4699-be90-a87db3943204 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.880249] env[63088]: INFO nova.virt.vmwareapi.vm_util [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Created folder: Project (f95a6d07d1ba4f618b3c6bd3960353fd) in parent group-v275816. [ 945.880587] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Creating folder: Instances. Parent ref: group-v275917. {{(pid=63088) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 945.880876] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-89185315-3af3-4d4b-823c-bb8b977d129d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.892036] env[63088]: INFO nova.virt.vmwareapi.vm_util [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Created folder: Instances in parent group-v275917. [ 945.892352] env[63088]: DEBUG oslo.service.loopingcall [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 945.892519] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8205f9e1-effd-442d-990b-bf5a2a9e6cdd] Creating VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 945.892761] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d57f5ec4-7471-4105-ab6d-f9c13b897c08 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.910605] env[63088]: DEBUG nova.compute.manager [req-be44789e-9518-4c93-9d6b-5bd329d61837 req-532100d9-534c-433b-ad1f-f259177e34a4 service nova] [instance: 8205f9e1-effd-442d-990b-bf5a2a9e6cdd] Received event network-changed-faac8e4c-2334-4c13-ae85-13e55b44fbb6 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 945.910833] env[63088]: DEBUG nova.compute.manager [req-be44789e-9518-4c93-9d6b-5bd329d61837 req-532100d9-534c-433b-ad1f-f259177e34a4 service nova] [instance: 8205f9e1-effd-442d-990b-bf5a2a9e6cdd] Refreshing instance network info cache due to event network-changed-faac8e4c-2334-4c13-ae85-13e55b44fbb6. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 945.911100] env[63088]: DEBUG oslo_concurrency.lockutils [req-be44789e-9518-4c93-9d6b-5bd329d61837 req-532100d9-534c-433b-ad1f-f259177e34a4 service nova] Acquiring lock "refresh_cache-8205f9e1-effd-442d-990b-bf5a2a9e6cdd" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 945.911284] env[63088]: DEBUG oslo_concurrency.lockutils [req-be44789e-9518-4c93-9d6b-5bd329d61837 req-532100d9-534c-433b-ad1f-f259177e34a4 service nova] Acquired lock "refresh_cache-8205f9e1-effd-442d-990b-bf5a2a9e6cdd" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 945.911486] env[63088]: DEBUG nova.network.neutron [req-be44789e-9518-4c93-9d6b-5bd329d61837 req-532100d9-534c-433b-ad1f-f259177e34a4 service nova] [instance: 8205f9e1-effd-442d-990b-bf5a2a9e6cdd] Refreshing network info cache for port faac8e4c-2334-4c13-ae85-13e55b44fbb6 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 945.919558] env[63088]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 945.919558] env[63088]: value = "task-1285242" [ 945.919558] env[63088]: _type = "Task" [ 945.919558] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.929972] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285242, 'name': CreateVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.968378] env[63088]: DEBUG oslo_vmware.api [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285238, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.013504] env[63088]: DEBUG oslo_vmware.api [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285239, 'name': Destroy_Task, 'duration_secs': 0.488424} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.013701] env[63088]: INFO nova.virt.vmwareapi.vm_util [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] Destroyed the VM [ 946.014382] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] Deleting Snapshot of the VM instance {{(pid=63088) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 946.015173] env[63088]: DEBUG nova.compute.manager [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 946.017824] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-759df3f4-6105-4444-8e00-01b15279df74 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.025367] env[63088]: DEBUG oslo_vmware.api [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 946.025367] env[63088]: value = "task-1285243" [ 946.025367] env[63088]: _type = "Task" [ 946.025367] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.033891] env[63088]: DEBUG oslo_vmware.api [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285243, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.042602] env[63088]: DEBUG nova.virt.hardware [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 946.042875] env[63088]: DEBUG nova.virt.hardware [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 946.043074] env[63088]: DEBUG nova.virt.hardware [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 946.043274] env[63088]: DEBUG nova.virt.hardware [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 946.043429] env[63088]: DEBUG nova.virt.hardware [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 946.043580] env[63088]: DEBUG nova.virt.hardware [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 946.043791] env[63088]: DEBUG nova.virt.hardware [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 946.043986] env[63088]: DEBUG nova.virt.hardware [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 946.044514] env[63088]: DEBUG nova.virt.hardware [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 946.044514] env[63088]: DEBUG nova.virt.hardware [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 946.044514] env[63088]: DEBUG nova.virt.hardware [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 946.045736] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db7a50e8-6530-4c16-96b2-2538b6c20ed0 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.055652] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fa7a993-f2ce-4d22-92af-a6509e1686e4 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.090732] env[63088]: DEBUG oslo_vmware.api [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285237, 'name': ReconfigVM_Task, 'duration_secs': 1.017947} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.091024] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2ad33254-8030-4454-b023-3359071f85de] Reconfigured VM instance instance-00000055 to attach disk [datastore2] 2ad33254-8030-4454-b023-3359071f85de/2ad33254-8030-4454-b023-3359071f85de.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 946.092027] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7bc8d945-a9ad-44fa-92d7-4758b4572eb7 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.098960] env[63088]: DEBUG oslo_vmware.api [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Waiting for the task: (returnval){ [ 946.098960] env[63088]: value = "task-1285244" [ 946.098960] env[63088]: _type = "Task" [ 946.098960] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.107852] env[63088]: DEBUG oslo_vmware.api [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285244, 'name': Rename_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.263398] env[63088]: DEBUG oslo_vmware.api [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285236, 'name': CopyVirtualDisk_Task} progress is 80%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.358027] env[63088]: DEBUG oslo_concurrency.lockutils [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.367s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.358027] env[63088]: DEBUG nova.compute.manager [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] [instance: 49ee4180-6322-4003-a5b1-f2a91d190290] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 946.360335] env[63088]: DEBUG oslo_concurrency.lockutils [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.723s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 946.362341] env[63088]: INFO nova.compute.claims [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] [instance: 4a56d74c-cb1f-4edf-99f0-4b54bf357a3a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 946.431684] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285242, 'name': CreateVM_Task} progress is 25%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.470956] env[63088]: DEBUG oslo_vmware.api [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285238, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.536262] env[63088]: DEBUG oslo_vmware.api [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285243, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.551766] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Acquiring lock "71e37d8e-a454-46c4-a3cc-3d5671a32beb" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 946.551766] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Lock "71e37d8e-a454-46c4-a3cc-3d5671a32beb" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 946.551766] env[63088]: INFO nova.compute.manager [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Shelving [ 946.568487] env[63088]: DEBUG nova.network.neutron [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Successfully updated port: b7a01048-9168-46ff-a5b8-83cee3dd3c97 {{(pid=63088) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 946.617026] env[63088]: DEBUG oslo_vmware.api [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285244, 'name': Rename_Task, 'duration_secs': 0.370822} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.617026] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2ad33254-8030-4454-b023-3359071f85de] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 946.618202] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ac8c1e79-74fa-4b7c-a019-6e1676153087 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.626721] env[63088]: DEBUG oslo_vmware.api [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Waiting for the task: (returnval){ [ 946.626721] env[63088]: value = "task-1285245" [ 946.626721] env[63088]: _type = "Task" [ 946.626721] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.636625] env[63088]: DEBUG oslo_vmware.api [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285245, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.766026] env[63088]: DEBUG oslo_vmware.api [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285236, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.444461} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.766549] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c89225f6-2d34-4988-ae47-d3d6903890da/c89225f6-2d34-4988-ae47-d3d6903890da.vmdk to [datastore2] 79c4d6e8-9999-4b0c-98d3-bbfd4d61e660/79c4d6e8-9999-4b0c-98d3-bbfd4d61e660.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 946.767600] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14d241b1-a1d4-470f-864a-491f75e71339 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.792517] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 79c4d6e8-9999-4b0c-98d3-bbfd4d61e660] Reconfiguring VM instance instance-00000054 to attach disk [datastore2] 79c4d6e8-9999-4b0c-98d3-bbfd4d61e660/79c4d6e8-9999-4b0c-98d3-bbfd4d61e660.vmdk or device None with type streamOptimized {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 946.793260] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6564a2c2-5bbb-4a71-92b3-547d2c279cbc {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.815391] env[63088]: DEBUG oslo_vmware.api [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Waiting for the task: (returnval){ [ 946.815391] env[63088]: value = "task-1285246" [ 946.815391] env[63088]: _type = "Task" [ 946.815391] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.825496] env[63088]: DEBUG oslo_vmware.api [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285246, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.869043] env[63088]: DEBUG nova.compute.utils [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 946.873038] env[63088]: DEBUG nova.compute.manager [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] [instance: 49ee4180-6322-4003-a5b1-f2a91d190290] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 946.873197] env[63088]: DEBUG nova.network.neutron [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] [instance: 49ee4180-6322-4003-a5b1-f2a91d190290] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 946.915113] env[63088]: DEBUG nova.policy [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd58472d1720f45c38cee1f250cd82339', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ef5fcb02afd04a42acbc92a0b89ea294', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 946.925456] env[63088]: DEBUG nova.network.neutron [req-be44789e-9518-4c93-9d6b-5bd329d61837 req-532100d9-534c-433b-ad1f-f259177e34a4 service nova] [instance: 8205f9e1-effd-442d-990b-bf5a2a9e6cdd] Updated VIF entry in instance network info cache for port faac8e4c-2334-4c13-ae85-13e55b44fbb6. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 946.925854] env[63088]: DEBUG nova.network.neutron [req-be44789e-9518-4c93-9d6b-5bd329d61837 req-532100d9-534c-433b-ad1f-f259177e34a4 service nova] [instance: 8205f9e1-effd-442d-990b-bf5a2a9e6cdd] Updating instance_info_cache with network_info: [{"id": "faac8e4c-2334-4c13-ae85-13e55b44fbb6", "address": "fa:16:3e:5f:f6:fa", "network": {"id": "cca88867-6df4-49c9-baec-a37aa8156a46", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-1660940561-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f95a6d07d1ba4f618b3c6bd3960353fd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4813d311-2016-4f6a-a4b8-7613ab624fba", "external-id": "nsx-vlan-transportzone-386", "segmentation_id": 386, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfaac8e4c-23", "ovs_interfaceid": "faac8e4c-2334-4c13-ae85-13e55b44fbb6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 946.933208] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285242, 'name': CreateVM_Task, 'duration_secs': 0.840475} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.933372] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8205f9e1-effd-442d-990b-bf5a2a9e6cdd] Created VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 946.934424] env[63088]: DEBUG oslo_concurrency.lockutils [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 946.934424] env[63088]: DEBUG oslo_concurrency.lockutils [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 946.934790] env[63088]: DEBUG oslo_concurrency.lockutils [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 946.935204] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-862f5504-8fb2-4464-b845-1a5300c9d500 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.940855] env[63088]: DEBUG oslo_vmware.api [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Waiting for the task: (returnval){ [ 946.940855] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]5211ae31-d6c8-67b4-6e7e-20eaa746f51e" [ 946.940855] env[63088]: _type = "Task" [ 946.940855] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.952932] env[63088]: DEBUG oslo_vmware.api [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]5211ae31-d6c8-67b4-6e7e-20eaa746f51e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.968856] env[63088]: DEBUG oslo_vmware.api [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285238, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.036028] env[63088]: DEBUG oslo_vmware.api [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285243, 'name': RemoveSnapshot_Task, 'duration_secs': 0.762474} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.036312] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] Deleted Snapshot of the VM instance {{(pid=63088) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 947.036600] env[63088]: DEBUG nova.compute.manager [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 947.037393] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f964b203-0969-4277-9eb3-dd8e01949563 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.058075] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 947.059060] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-18cc74b7-8973-4034-85d8-d632aa1efb78 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.064853] env[63088]: DEBUG oslo_vmware.api [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Waiting for the task: (returnval){ [ 947.064853] env[63088]: value = "task-1285247" [ 947.064853] env[63088]: _type = "Task" [ 947.064853] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.073519] env[63088]: DEBUG oslo_concurrency.lockutils [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquiring lock "refresh_cache-d6fef3ef-fec8-4929-b9b8-5e63306aa51d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 947.073763] env[63088]: DEBUG oslo_concurrency.lockutils [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquired lock "refresh_cache-d6fef3ef-fec8-4929-b9b8-5e63306aa51d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 947.073858] env[63088]: DEBUG nova.network.neutron [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 947.075508] env[63088]: DEBUG oslo_vmware.api [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285247, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.137548] env[63088]: DEBUG oslo_vmware.api [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285245, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.179286] env[63088]: DEBUG nova.network.neutron [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] [instance: 49ee4180-6322-4003-a5b1-f2a91d190290] Successfully created port: 15b6a284-5552-4594-97bf-23b7ec018f5d {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 947.326565] env[63088]: DEBUG oslo_vmware.api [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285246, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.377026] env[63088]: DEBUG nova.compute.manager [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] [instance: 49ee4180-6322-4003-a5b1-f2a91d190290] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 947.429112] env[63088]: DEBUG oslo_concurrency.lockutils [req-be44789e-9518-4c93-9d6b-5bd329d61837 req-532100d9-534c-433b-ad1f-f259177e34a4 service nova] Releasing lock "refresh_cache-8205f9e1-effd-442d-990b-bf5a2a9e6cdd" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 947.457058] env[63088]: DEBUG oslo_vmware.api [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]5211ae31-d6c8-67b4-6e7e-20eaa746f51e, 'name': SearchDatastore_Task, 'duration_secs': 0.05061} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.459133] env[63088]: DEBUG oslo_concurrency.lockutils [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 947.459408] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] [instance: 8205f9e1-effd-442d-990b-bf5a2a9e6cdd] Processing image 1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 947.459654] env[63088]: DEBUG oslo_concurrency.lockutils [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 947.459806] env[63088]: DEBUG oslo_concurrency.lockutils [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 947.459994] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 947.460378] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d141edeb-78a4-41b1-9145-9229d4ae6141 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.474341] env[63088]: DEBUG oslo_vmware.api [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285238, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.052675} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.474627] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore2] 7cd45e52-c6be-4ec6-9710-fa99e29cf872/7cd45e52-c6be-4ec6-9710-fa99e29cf872.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 947.474816] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 7cd45e52-c6be-4ec6-9710-fa99e29cf872] Extending root virtual disk to 1048576 {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 947.476054] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3fc9a149-2f2f-4319-98a1-6f6c8a1962be {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.478227] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 947.478227] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 947.479224] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f4ff0500-0b9e-4b87-b721-bd84b790a690 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.490424] env[63088]: DEBUG oslo_vmware.api [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Waiting for the task: (returnval){ [ 947.490424] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]5209a7cf-079d-d8e2-a43c-f57bdedbe4b2" [ 947.490424] env[63088]: _type = "Task" [ 947.490424] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.490702] env[63088]: DEBUG oslo_vmware.api [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Waiting for the task: (returnval){ [ 947.490702] env[63088]: value = "task-1285248" [ 947.490702] env[63088]: _type = "Task" [ 947.490702] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.506440] env[63088]: DEBUG oslo_vmware.api [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285248, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.506440] env[63088]: DEBUG oslo_vmware.api [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]5209a7cf-079d-d8e2-a43c-f57bdedbe4b2, 'name': SearchDatastore_Task, 'duration_secs': 0.008951} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.509022] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7cfd13bc-b83a-4c51-be40-ae87ad617efa {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.512469] env[63088]: DEBUG oslo_vmware.api [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Waiting for the task: (returnval){ [ 947.512469] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]525d299d-1ac2-4e8e-a2e7-935ce00c8273" [ 947.512469] env[63088]: _type = "Task" [ 947.512469] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.523845] env[63088]: DEBUG oslo_vmware.api [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]525d299d-1ac2-4e8e-a2e7-935ce00c8273, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.551456] env[63088]: INFO nova.compute.manager [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] Shelve offloading [ 947.552790] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 947.553052] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-756d126c-3f14-4566-9bc3-2c93895a261e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.559442] env[63088]: DEBUG oslo_vmware.api [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 947.559442] env[63088]: value = "task-1285249" [ 947.559442] env[63088]: _type = "Task" [ 947.559442] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.572178] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] VM already powered off {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 947.572381] env[63088]: DEBUG nova.compute.manager [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 947.573143] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d367cdb7-1a4e-4a81-8a52-c34944782fb9 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.582070] env[63088]: DEBUG oslo_vmware.api [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285247, 'name': PowerOffVM_Task, 'duration_secs': 0.482812} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.585034] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 947.586675] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1676d843-a5ef-4f53-8e74-16f4a19ca2f0 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.590742] env[63088]: DEBUG oslo_concurrency.lockutils [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquiring lock "refresh_cache-72c07d1e-cbb6-4875-9b7f-2649aa25eee5" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 947.590936] env[63088]: DEBUG oslo_concurrency.lockutils [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquired lock "refresh_cache-72c07d1e-cbb6-4875-9b7f-2649aa25eee5" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 947.591153] env[63088]: DEBUG nova.network.neutron [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 947.612423] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-211e0a66-6026-4de9-933e-0d6b38271dae {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.640754] env[63088]: DEBUG oslo_vmware.api [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285245, 'name': PowerOnVM_Task, 'duration_secs': 0.698417} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.641068] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2ad33254-8030-4454-b023-3359071f85de] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 947.641603] env[63088]: INFO nova.compute.manager [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2ad33254-8030-4454-b023-3359071f85de] Took 10.55 seconds to spawn the instance on the hypervisor. [ 947.641603] env[63088]: DEBUG nova.compute.manager [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2ad33254-8030-4454-b023-3359071f85de] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 947.643079] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-444a7a14-8dbe-4484-88d2-44097492bb65 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.654238] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-955754f6-ff70-465e-b774-8ff212f48d27 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.661708] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d67192f4-b1f8-4915-a2e4-8788a40bda79 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.667876] env[63088]: DEBUG nova.network.neutron [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 947.700775] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6934b1d9-592d-4583-a2f3-4b5b1f193d12 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.709622] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40261b7b-14d2-4a8c-af94-b5d36b4238b4 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.723661] env[63088]: DEBUG nova.compute.provider_tree [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 947.828975] env[63088]: DEBUG oslo_vmware.api [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285246, 'name': ReconfigVM_Task, 'duration_secs': 0.530445} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.829313] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 79c4d6e8-9999-4b0c-98d3-bbfd4d61e660] Reconfigured VM instance instance-00000054 to attach disk [datastore2] 79c4d6e8-9999-4b0c-98d3-bbfd4d61e660/79c4d6e8-9999-4b0c-98d3-bbfd4d61e660.vmdk or device None with type streamOptimized {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 947.829950] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8d76a0fb-a281-4b2b-94ad-ab04cc9da730 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.836561] env[63088]: DEBUG oslo_vmware.api [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Waiting for the task: (returnval){ [ 947.836561] env[63088]: value = "task-1285250" [ 947.836561] env[63088]: _type = "Task" [ 947.836561] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.844320] env[63088]: DEBUG oslo_vmware.api [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285250, 'name': Rename_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.990491] env[63088]: DEBUG nova.compute.manager [req-6d4f6ef5-63ac-4be1-bb1d-ad67490b571a req-5fb95b22-f607-4e42-8ca2-b3e293db5a9d service nova] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Received event network-vif-plugged-b7a01048-9168-46ff-a5b8-83cee3dd3c97 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 947.990768] env[63088]: DEBUG oslo_concurrency.lockutils [req-6d4f6ef5-63ac-4be1-bb1d-ad67490b571a req-5fb95b22-f607-4e42-8ca2-b3e293db5a9d service nova] Acquiring lock "d6fef3ef-fec8-4929-b9b8-5e63306aa51d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.990943] env[63088]: DEBUG oslo_concurrency.lockutils [req-6d4f6ef5-63ac-4be1-bb1d-ad67490b571a req-5fb95b22-f607-4e42-8ca2-b3e293db5a9d service nova] Lock "d6fef3ef-fec8-4929-b9b8-5e63306aa51d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.991248] env[63088]: DEBUG oslo_concurrency.lockutils [req-6d4f6ef5-63ac-4be1-bb1d-ad67490b571a req-5fb95b22-f607-4e42-8ca2-b3e293db5a9d service nova] Lock "d6fef3ef-fec8-4929-b9b8-5e63306aa51d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.991354] env[63088]: DEBUG nova.compute.manager [req-6d4f6ef5-63ac-4be1-bb1d-ad67490b571a req-5fb95b22-f607-4e42-8ca2-b3e293db5a9d service nova] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] No waiting events found dispatching network-vif-plugged-b7a01048-9168-46ff-a5b8-83cee3dd3c97 {{(pid=63088) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 947.991587] env[63088]: WARNING nova.compute.manager [req-6d4f6ef5-63ac-4be1-bb1d-ad67490b571a req-5fb95b22-f607-4e42-8ca2-b3e293db5a9d service nova] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Received unexpected event network-vif-plugged-b7a01048-9168-46ff-a5b8-83cee3dd3c97 for instance with vm_state building and task_state spawning. [ 947.991731] env[63088]: DEBUG nova.compute.manager [req-6d4f6ef5-63ac-4be1-bb1d-ad67490b571a req-5fb95b22-f607-4e42-8ca2-b3e293db5a9d service nova] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Received event network-changed-b7a01048-9168-46ff-a5b8-83cee3dd3c97 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 947.991928] env[63088]: DEBUG nova.compute.manager [req-6d4f6ef5-63ac-4be1-bb1d-ad67490b571a req-5fb95b22-f607-4e42-8ca2-b3e293db5a9d service nova] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Refreshing instance network info cache due to event network-changed-b7a01048-9168-46ff-a5b8-83cee3dd3c97. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 947.992122] env[63088]: DEBUG oslo_concurrency.lockutils [req-6d4f6ef5-63ac-4be1-bb1d-ad67490b571a req-5fb95b22-f607-4e42-8ca2-b3e293db5a9d service nova] Acquiring lock "refresh_cache-d6fef3ef-fec8-4929-b9b8-5e63306aa51d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 948.001730] env[63088]: DEBUG oslo_vmware.api [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285248, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.317562} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.001986] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 7cd45e52-c6be-4ec6-9710-fa99e29cf872] Extended root virtual disk {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 948.002809] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80e5bca9-efb3-418b-bd8b-94f912f930bf {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.026376] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 7cd45e52-c6be-4ec6-9710-fa99e29cf872] Reconfiguring VM instance instance-00000056 to attach disk [datastore2] 7cd45e52-c6be-4ec6-9710-fa99e29cf872/7cd45e52-c6be-4ec6-9710-fa99e29cf872.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 948.029925] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c417e2d0-97b2-4dea-9f00-6bafa9e1590f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.045945] env[63088]: DEBUG nova.network.neutron [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Updating instance_info_cache with network_info: [{"id": "b7a01048-9168-46ff-a5b8-83cee3dd3c97", "address": "fa:16:3e:e3:57:91", "network": {"id": "1174a284-d3ca-4f9e-aa81-13ee9a693e55", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1994276040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa3d24a1a6c0430985fd80365d986ee1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb7a01048-91", "ovs_interfaceid": "b7a01048-9168-46ff-a5b8-83cee3dd3c97", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 948.053962] env[63088]: DEBUG oslo_vmware.api [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]525d299d-1ac2-4e8e-a2e7-935ce00c8273, 'name': SearchDatastore_Task, 'duration_secs': 0.010519} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.055321] env[63088]: DEBUG oslo_concurrency.lockutils [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 948.055592] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore2] 8205f9e1-effd-442d-990b-bf5a2a9e6cdd/8205f9e1-effd-442d-990b-bf5a2a9e6cdd.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 948.055940] env[63088]: DEBUG oslo_vmware.api [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Waiting for the task: (returnval){ [ 948.055940] env[63088]: value = "task-1285251" [ 948.055940] env[63088]: _type = "Task" [ 948.055940] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.056192] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0309dd9f-ae70-4fea-9fc7-f2e0151c24c6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.067750] env[63088]: DEBUG oslo_vmware.api [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285251, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.068953] env[63088]: DEBUG oslo_vmware.api [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Waiting for the task: (returnval){ [ 948.068953] env[63088]: value = "task-1285252" [ 948.068953] env[63088]: _type = "Task" [ 948.068953] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.077974] env[63088]: DEBUG oslo_vmware.api [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Task: {'id': task-1285252, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.123850] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Creating Snapshot of the VM instance {{(pid=63088) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 948.124276] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-c6063f7d-a5d2-4ac5-b95b-f5f1e89fedab {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.132986] env[63088]: DEBUG oslo_vmware.api [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Waiting for the task: (returnval){ [ 948.132986] env[63088]: value = "task-1285253" [ 948.132986] env[63088]: _type = "Task" [ 948.132986] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.143719] env[63088]: DEBUG oslo_vmware.api [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285253, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.165020] env[63088]: INFO nova.compute.manager [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2ad33254-8030-4454-b023-3359071f85de] Took 26.14 seconds to build instance. [ 948.226677] env[63088]: DEBUG nova.scheduler.client.report [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 948.366791] env[63088]: DEBUG oslo_vmware.api [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285250, 'name': Rename_Task, 'duration_secs': 0.313615} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.367158] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 79c4d6e8-9999-4b0c-98d3-bbfd4d61e660] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 948.367441] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c4a2fc34-a072-4487-963c-3ce1d382d9ad {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.375616] env[63088]: DEBUG oslo_vmware.api [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Waiting for the task: (returnval){ [ 948.375616] env[63088]: value = "task-1285254" [ 948.375616] env[63088]: _type = "Task" [ 948.375616] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.392863] env[63088]: DEBUG nova.compute.manager [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] [instance: 49ee4180-6322-4003-a5b1-f2a91d190290] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 948.396083] env[63088]: DEBUG oslo_vmware.api [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285254, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.436547] env[63088]: DEBUG nova.virt.hardware [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 948.437146] env[63088]: DEBUG nova.virt.hardware [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 948.437307] env[63088]: DEBUG nova.virt.hardware [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 948.437509] env[63088]: DEBUG nova.virt.hardware [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 948.437663] env[63088]: DEBUG nova.virt.hardware [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 948.437820] env[63088]: DEBUG nova.virt.hardware [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 948.438160] env[63088]: DEBUG nova.virt.hardware [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 948.438527] env[63088]: DEBUG nova.virt.hardware [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 948.438836] env[63088]: DEBUG nova.virt.hardware [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 948.439078] env[63088]: DEBUG nova.virt.hardware [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 948.439320] env[63088]: DEBUG nova.virt.hardware [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 948.440362] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3be52912-d64e-492e-8455-9a8971a1b6d0 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.450300] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b22ba87a-ce9a-41d8-bc3b-a541ca3c4df3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.520236] env[63088]: DEBUG nova.network.neutron [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] Updating instance_info_cache with network_info: [{"id": "88203cff-5be6-4261-a168-197e6be13cc3", "address": "fa:16:3e:e8:83:1f", "network": {"id": "5b801381-6bb3-45cd-b81f-92b5d757e080", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1481457494-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e48f62dee9ad4e9b94b67bd871db5c63", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "051f343d-ac4f-4070-a26d-467603122c81", "external-id": "nsx-vlan-transportzone-277", "segmentation_id": 277, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap88203cff-5b", "ovs_interfaceid": "88203cff-5be6-4261-a168-197e6be13cc3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 948.549024] env[63088]: DEBUG oslo_concurrency.lockutils [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Releasing lock "refresh_cache-d6fef3ef-fec8-4929-b9b8-5e63306aa51d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 948.549887] env[63088]: DEBUG nova.compute.manager [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Instance network_info: |[{"id": "b7a01048-9168-46ff-a5b8-83cee3dd3c97", "address": "fa:16:3e:e3:57:91", "network": {"id": "1174a284-d3ca-4f9e-aa81-13ee9a693e55", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1994276040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa3d24a1a6c0430985fd80365d986ee1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb7a01048-91", "ovs_interfaceid": "b7a01048-9168-46ff-a5b8-83cee3dd3c97", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 948.550249] env[63088]: DEBUG oslo_concurrency.lockutils [req-6d4f6ef5-63ac-4be1-bb1d-ad67490b571a req-5fb95b22-f607-4e42-8ca2-b3e293db5a9d service nova] Acquired lock "refresh_cache-d6fef3ef-fec8-4929-b9b8-5e63306aa51d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 948.550510] env[63088]: DEBUG nova.network.neutron [req-6d4f6ef5-63ac-4be1-bb1d-ad67490b571a req-5fb95b22-f607-4e42-8ca2-b3e293db5a9d service nova] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Refreshing network info cache for port b7a01048-9168-46ff-a5b8-83cee3dd3c97 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 948.551818] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e3:57:91', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b7a01048-9168-46ff-a5b8-83cee3dd3c97', 'vif_model': 'vmxnet3'}] {{(pid=63088) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 948.562296] env[63088]: DEBUG oslo.service.loopingcall [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 948.562296] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Creating VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 948.565938] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-03cbe665-da8c-4b87-bdec-d02d4db69109 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.590983] env[63088]: DEBUG oslo_vmware.api [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285251, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.592901] env[63088]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 948.592901] env[63088]: value = "task-1285255" [ 948.592901] env[63088]: _type = "Task" [ 948.592901] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.602518] env[63088]: DEBUG oslo_vmware.api [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Task: {'id': task-1285252, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.500242} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.602518] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore2] 8205f9e1-effd-442d-990b-bf5a2a9e6cdd/8205f9e1-effd-442d-990b-bf5a2a9e6cdd.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 948.602518] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] [instance: 8205f9e1-effd-442d-990b-bf5a2a9e6cdd] Extending root virtual disk to 1048576 {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 948.602518] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-da5151f7-e2ab-461e-ad75-0da473d85cf5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.608854] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285255, 'name': CreateVM_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.614558] env[63088]: DEBUG oslo_vmware.api [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Waiting for the task: (returnval){ [ 948.614558] env[63088]: value = "task-1285256" [ 948.614558] env[63088]: _type = "Task" [ 948.614558] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.630077] env[63088]: DEBUG oslo_vmware.api [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Task: {'id': task-1285256, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.643276] env[63088]: DEBUG oslo_vmware.api [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285253, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.665871] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d5a4d720-7f1b-48b8-9625-fd546c8f5178 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Lock "2ad33254-8030-4454-b023-3359071f85de" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.646s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 948.732503] env[63088]: DEBUG oslo_concurrency.lockutils [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.373s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 948.733088] env[63088]: DEBUG nova.compute.manager [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] [instance: 4a56d74c-cb1f-4edf-99f0-4b54bf357a3a] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 948.884729] env[63088]: DEBUG nova.compute.manager [req-9dd6a658-79c2-4829-8acf-e58790debd6e req-6555003a-c6dd-4240-b81b-9a1ba5511be8 service nova] [instance: 49ee4180-6322-4003-a5b1-f2a91d190290] Received event network-vif-plugged-15b6a284-5552-4594-97bf-23b7ec018f5d {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 948.884846] env[63088]: DEBUG oslo_concurrency.lockutils [req-9dd6a658-79c2-4829-8acf-e58790debd6e req-6555003a-c6dd-4240-b81b-9a1ba5511be8 service nova] Acquiring lock "49ee4180-6322-4003-a5b1-f2a91d190290-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 948.885049] env[63088]: DEBUG oslo_concurrency.lockutils [req-9dd6a658-79c2-4829-8acf-e58790debd6e req-6555003a-c6dd-4240-b81b-9a1ba5511be8 service nova] Lock "49ee4180-6322-4003-a5b1-f2a91d190290-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 948.885139] env[63088]: DEBUG oslo_concurrency.lockutils [req-9dd6a658-79c2-4829-8acf-e58790debd6e req-6555003a-c6dd-4240-b81b-9a1ba5511be8 service nova] Lock "49ee4180-6322-4003-a5b1-f2a91d190290-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 948.885856] env[63088]: DEBUG nova.compute.manager [req-9dd6a658-79c2-4829-8acf-e58790debd6e req-6555003a-c6dd-4240-b81b-9a1ba5511be8 service nova] [instance: 49ee4180-6322-4003-a5b1-f2a91d190290] No waiting events found dispatching network-vif-plugged-15b6a284-5552-4594-97bf-23b7ec018f5d {{(pid=63088) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 948.886424] env[63088]: WARNING nova.compute.manager [req-9dd6a658-79c2-4829-8acf-e58790debd6e req-6555003a-c6dd-4240-b81b-9a1ba5511be8 service nova] [instance: 49ee4180-6322-4003-a5b1-f2a91d190290] Received unexpected event network-vif-plugged-15b6a284-5552-4594-97bf-23b7ec018f5d for instance with vm_state building and task_state spawning. [ 948.893694] env[63088]: DEBUG oslo_vmware.api [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285254, 'name': PowerOnVM_Task} progress is 87%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.023875] env[63088]: DEBUG oslo_concurrency.lockutils [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Releasing lock "refresh_cache-72c07d1e-cbb6-4875-9b7f-2649aa25eee5" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 949.075277] env[63088]: DEBUG oslo_vmware.api [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285251, 'name': ReconfigVM_Task, 'duration_secs': 0.612562} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.075619] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 7cd45e52-c6be-4ec6-9710-fa99e29cf872] Reconfigured VM instance instance-00000056 to attach disk [datastore2] 7cd45e52-c6be-4ec6-9710-fa99e29cf872/7cd45e52-c6be-4ec6-9710-fa99e29cf872.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 949.076309] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4944ace9-c99e-4f78-aee5-5c4ab7c2f9da {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.082230] env[63088]: DEBUG oslo_vmware.api [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Waiting for the task: (returnval){ [ 949.082230] env[63088]: value = "task-1285257" [ 949.082230] env[63088]: _type = "Task" [ 949.082230] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.105742] env[63088]: DEBUG oslo_vmware.api [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285257, 'name': Rename_Task} progress is 10%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.111968] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285255, 'name': CreateVM_Task} progress is 99%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.124441] env[63088]: DEBUG oslo_vmware.api [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Task: {'id': task-1285256, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.104836} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.124956] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] [instance: 8205f9e1-effd-442d-990b-bf5a2a9e6cdd] Extended root virtual disk {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 949.129017] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29286953-6bad-4036-bc83-200288262fec {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.149330] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] [instance: 8205f9e1-effd-442d-990b-bf5a2a9e6cdd] Reconfiguring VM instance instance-00000057 to attach disk [datastore2] 8205f9e1-effd-442d-990b-bf5a2a9e6cdd/8205f9e1-effd-442d-990b-bf5a2a9e6cdd.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 949.155119] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-55530a73-69f5-43e1-8498-20bcbf35e0d4 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.177465] env[63088]: DEBUG oslo_vmware.api [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285253, 'name': CreateSnapshot_Task, 'duration_secs': 0.817103} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.178785] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Created Snapshot of the VM instance {{(pid=63088) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 949.179140] env[63088]: DEBUG oslo_vmware.api [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Waiting for the task: (returnval){ [ 949.179140] env[63088]: value = "task-1285258" [ 949.179140] env[63088]: _type = "Task" [ 949.179140] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.179840] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d33d3b5-c805-419e-bf91-84951cb9affd {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.195435] env[63088]: DEBUG oslo_vmware.api [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Task: {'id': task-1285258, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.239746] env[63088]: DEBUG nova.compute.utils [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 949.241591] env[63088]: DEBUG nova.compute.manager [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] [instance: 4a56d74c-cb1f-4edf-99f0-4b54bf357a3a] Not allocating networking since 'none' was specified. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 949.342356] env[63088]: DEBUG nova.compute.manager [req-48dd4b6f-ff5f-4530-8ce2-750a823158a9 req-25d44115-bb2d-4ccc-9e33-34196b6117af service nova] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] Received event network-vif-unplugged-88203cff-5be6-4261-a168-197e6be13cc3 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 949.342412] env[63088]: DEBUG oslo_concurrency.lockutils [req-48dd4b6f-ff5f-4530-8ce2-750a823158a9 req-25d44115-bb2d-4ccc-9e33-34196b6117af service nova] Acquiring lock "72c07d1e-cbb6-4875-9b7f-2649aa25eee5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.342637] env[63088]: DEBUG oslo_concurrency.lockutils [req-48dd4b6f-ff5f-4530-8ce2-750a823158a9 req-25d44115-bb2d-4ccc-9e33-34196b6117af service nova] Lock "72c07d1e-cbb6-4875-9b7f-2649aa25eee5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.342777] env[63088]: DEBUG oslo_concurrency.lockutils [req-48dd4b6f-ff5f-4530-8ce2-750a823158a9 req-25d44115-bb2d-4ccc-9e33-34196b6117af service nova] Lock "72c07d1e-cbb6-4875-9b7f-2649aa25eee5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.342957] env[63088]: DEBUG nova.compute.manager [req-48dd4b6f-ff5f-4530-8ce2-750a823158a9 req-25d44115-bb2d-4ccc-9e33-34196b6117af service nova] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] No waiting events found dispatching network-vif-unplugged-88203cff-5be6-4261-a168-197e6be13cc3 {{(pid=63088) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 949.343601] env[63088]: WARNING nova.compute.manager [req-48dd4b6f-ff5f-4530-8ce2-750a823158a9 req-25d44115-bb2d-4ccc-9e33-34196b6117af service nova] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] Received unexpected event network-vif-unplugged-88203cff-5be6-4261-a168-197e6be13cc3 for instance with vm_state shelved and task_state shelving_offloading. [ 949.368459] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 949.369518] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d48dd751-a998-404f-9cef-745ddf1b6558 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.378183] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] Unregistering the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 949.382322] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cd2c3717-ee5b-468d-9e42-b4f3c67e41e9 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.388868] env[63088]: DEBUG oslo_vmware.api [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285254, 'name': PowerOnVM_Task, 'duration_secs': 0.856895} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.389214] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 79c4d6e8-9999-4b0c-98d3-bbfd4d61e660] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 949.389442] env[63088]: INFO nova.compute.manager [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 79c4d6e8-9999-4b0c-98d3-bbfd4d61e660] Took 14.65 seconds to spawn the instance on the hypervisor. [ 949.389517] env[63088]: DEBUG nova.compute.manager [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 79c4d6e8-9999-4b0c-98d3-bbfd4d61e660] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 949.390293] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f8ea82b-d7c1-49f7-a21b-e81247c8e8e7 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.395149] env[63088]: DEBUG nova.network.neutron [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] [instance: 49ee4180-6322-4003-a5b1-f2a91d190290] Successfully updated port: 15b6a284-5552-4594-97bf-23b7ec018f5d {{(pid=63088) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 949.404838] env[63088]: DEBUG nova.network.neutron [req-6d4f6ef5-63ac-4be1-bb1d-ad67490b571a req-5fb95b22-f607-4e42-8ca2-b3e293db5a9d service nova] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Updated VIF entry in instance network info cache for port b7a01048-9168-46ff-a5b8-83cee3dd3c97. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 949.405179] env[63088]: DEBUG nova.network.neutron [req-6d4f6ef5-63ac-4be1-bb1d-ad67490b571a req-5fb95b22-f607-4e42-8ca2-b3e293db5a9d service nova] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Updating instance_info_cache with network_info: [{"id": "b7a01048-9168-46ff-a5b8-83cee3dd3c97", "address": "fa:16:3e:e3:57:91", "network": {"id": "1174a284-d3ca-4f9e-aa81-13ee9a693e55", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1994276040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa3d24a1a6c0430985fd80365d986ee1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb7a01048-91", "ovs_interfaceid": "b7a01048-9168-46ff-a5b8-83cee3dd3c97", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 949.445064] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] Unregistered the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 949.445313] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] Deleting contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 949.445495] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Deleting the datastore file [datastore1] 72c07d1e-cbb6-4875-9b7f-2649aa25eee5 {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 949.446048] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c6a0874f-6207-483e-825d-be8212019d99 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.451603] env[63088]: DEBUG oslo_vmware.api [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 949.451603] env[63088]: value = "task-1285260" [ 949.451603] env[63088]: _type = "Task" [ 949.451603] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.461791] env[63088]: DEBUG oslo_vmware.api [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285260, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.592235] env[63088]: DEBUG oslo_vmware.api [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285257, 'name': Rename_Task, 'duration_secs': 0.137603} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.592604] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 7cd45e52-c6be-4ec6-9710-fa99e29cf872] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 949.592821] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1b302cdc-1642-4a64-a68c-b82c5dbd7df5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.601577] env[63088]: DEBUG nova.compute.manager [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2ad33254-8030-4454-b023-3359071f85de] Stashing vm_state: active {{(pid=63088) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 949.611070] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285255, 'name': CreateVM_Task, 'duration_secs': 0.544738} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.611070] env[63088]: DEBUG oslo_vmware.api [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Waiting for the task: (returnval){ [ 949.611070] env[63088]: value = "task-1285261" [ 949.611070] env[63088]: _type = "Task" [ 949.611070] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.611070] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Created VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 949.611070] env[63088]: DEBUG oslo_concurrency.lockutils [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 949.611070] env[63088]: DEBUG oslo_concurrency.lockutils [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 949.611070] env[63088]: DEBUG oslo_concurrency.lockutils [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 949.613980] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ff3af32d-7c6f-481b-af5b-7e298b080040 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.621132] env[63088]: DEBUG oslo_vmware.api [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285261, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.622530] env[63088]: DEBUG oslo_vmware.api [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Waiting for the task: (returnval){ [ 949.622530] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]5263a439-c8ac-ea23-1abb-37d8305e37fb" [ 949.622530] env[63088]: _type = "Task" [ 949.622530] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.632331] env[63088]: DEBUG oslo_vmware.api [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]5263a439-c8ac-ea23-1abb-37d8305e37fb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.692263] env[63088]: DEBUG oslo_vmware.api [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Task: {'id': task-1285258, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.708840] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Creating linked-clone VM from snapshot {{(pid=63088) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 949.709170] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-bfc63295-990a-4b48-98ae-9d385a184721 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.717707] env[63088]: DEBUG oslo_vmware.api [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Waiting for the task: (returnval){ [ 949.717707] env[63088]: value = "task-1285262" [ 949.717707] env[63088]: _type = "Task" [ 949.717707] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.725891] env[63088]: DEBUG oslo_vmware.api [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285262, 'name': CloneVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.742603] env[63088]: DEBUG nova.compute.manager [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] [instance: 4a56d74c-cb1f-4edf-99f0-4b54bf357a3a] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 949.897705] env[63088]: DEBUG oslo_concurrency.lockutils [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Acquiring lock "refresh_cache-49ee4180-6322-4003-a5b1-f2a91d190290" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 949.897877] env[63088]: DEBUG oslo_concurrency.lockutils [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Acquired lock "refresh_cache-49ee4180-6322-4003-a5b1-f2a91d190290" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 949.898044] env[63088]: DEBUG nova.network.neutron [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] [instance: 49ee4180-6322-4003-a5b1-f2a91d190290] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 949.913945] env[63088]: DEBUG oslo_concurrency.lockutils [req-6d4f6ef5-63ac-4be1-bb1d-ad67490b571a req-5fb95b22-f607-4e42-8ca2-b3e293db5a9d service nova] Releasing lock "refresh_cache-d6fef3ef-fec8-4929-b9b8-5e63306aa51d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 949.917477] env[63088]: INFO nova.compute.manager [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 79c4d6e8-9999-4b0c-98d3-bbfd4d61e660] Took 30.09 seconds to build instance. [ 949.962344] env[63088]: DEBUG oslo_vmware.api [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285260, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.171698} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.962738] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 949.962993] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] Deleted contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 949.963235] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 949.988459] env[63088]: INFO nova.scheduler.client.report [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Deleted allocations for instance 72c07d1e-cbb6-4875-9b7f-2649aa25eee5 [ 950.057289] env[63088]: DEBUG nova.compute.manager [req-ee1c8620-275a-4e5c-b0d8-ba3750eb22dc req-c32e500a-e53a-4f65-8240-78841d878bb4 service nova] [instance: 49ee4180-6322-4003-a5b1-f2a91d190290] Received event network-changed-15b6a284-5552-4594-97bf-23b7ec018f5d {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 950.057289] env[63088]: DEBUG nova.compute.manager [req-ee1c8620-275a-4e5c-b0d8-ba3750eb22dc req-c32e500a-e53a-4f65-8240-78841d878bb4 service nova] [instance: 49ee4180-6322-4003-a5b1-f2a91d190290] Refreshing instance network info cache due to event network-changed-15b6a284-5552-4594-97bf-23b7ec018f5d. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 950.057421] env[63088]: DEBUG oslo_concurrency.lockutils [req-ee1c8620-275a-4e5c-b0d8-ba3750eb22dc req-c32e500a-e53a-4f65-8240-78841d878bb4 service nova] Acquiring lock "refresh_cache-49ee4180-6322-4003-a5b1-f2a91d190290" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 950.123806] env[63088]: DEBUG oslo_vmware.api [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285261, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.124781] env[63088]: DEBUG oslo_concurrency.lockutils [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 950.125048] env[63088]: DEBUG oslo_concurrency.lockutils [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 950.140883] env[63088]: DEBUG oslo_vmware.api [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]5263a439-c8ac-ea23-1abb-37d8305e37fb, 'name': SearchDatastore_Task, 'duration_secs': 0.015295} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.142283] env[63088]: DEBUG oslo_concurrency.lockutils [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 950.142802] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Processing image 1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 950.143066] env[63088]: DEBUG oslo_concurrency.lockutils [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 950.143779] env[63088]: DEBUG oslo_concurrency.lockutils [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 950.143779] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 950.144078] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-abc99811-54a2-4498-b46b-f532d7f30317 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.156180] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 950.156485] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 950.157799] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c7579503-c173-43e6-9622-06e490afb679 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.165883] env[63088]: DEBUG oslo_vmware.api [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Waiting for the task: (returnval){ [ 950.165883] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52e89bd6-cc5c-dd91-047a-ff4ea55f0cf7" [ 950.165883] env[63088]: _type = "Task" [ 950.165883] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.176121] env[63088]: DEBUG oslo_vmware.api [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52e89bd6-cc5c-dd91-047a-ff4ea55f0cf7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.192344] env[63088]: DEBUG oslo_vmware.api [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Task: {'id': task-1285258, 'name': ReconfigVM_Task, 'duration_secs': 0.716877} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.192643] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] [instance: 8205f9e1-effd-442d-990b-bf5a2a9e6cdd] Reconfigured VM instance instance-00000057 to attach disk [datastore2] 8205f9e1-effd-442d-990b-bf5a2a9e6cdd/8205f9e1-effd-442d-990b-bf5a2a9e6cdd.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 950.193321] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-93126e51-ecfe-45c7-b476-833bf5221fc4 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.199374] env[63088]: DEBUG oslo_vmware.api [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Waiting for the task: (returnval){ [ 950.199374] env[63088]: value = "task-1285263" [ 950.199374] env[63088]: _type = "Task" [ 950.199374] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.207462] env[63088]: DEBUG oslo_vmware.api [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Task: {'id': task-1285263, 'name': Rename_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.227418] env[63088]: DEBUG oslo_vmware.api [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285262, 'name': CloneVM_Task} progress is 94%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.420525] env[63088]: DEBUG oslo_concurrency.lockutils [None req-51cfdadf-98e7-43a0-ba3e-a00ecdc9dd61 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Lock "79c4d6e8-9999-4b0c-98d3-bbfd4d61e660" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.605s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.435082] env[63088]: DEBUG nova.network.neutron [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] [instance: 49ee4180-6322-4003-a5b1-f2a91d190290] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 950.494874] env[63088]: DEBUG oslo_concurrency.lockutils [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 950.565117] env[63088]: DEBUG nova.network.neutron [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] [instance: 49ee4180-6322-4003-a5b1-f2a91d190290] Updating instance_info_cache with network_info: [{"id": "15b6a284-5552-4594-97bf-23b7ec018f5d", "address": "fa:16:3e:2f:58:37", "network": {"id": "b16c35d8-18ac-44b6-9aed-4e3a48d8b6d0", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-1225586833-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ef5fcb02afd04a42acbc92a0b89ea294", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "42f08482-a1da-405d-9918-d733d9f5173c", "external-id": "nsx-vlan-transportzone-381", "segmentation_id": 381, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap15b6a284-55", "ovs_interfaceid": "15b6a284-5552-4594-97bf-23b7ec018f5d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 950.606520] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0dd094f1-4e17-4137-b0a9-4224c597f8a7 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Acquiring lock "79c4d6e8-9999-4b0c-98d3-bbfd4d61e660" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 950.606798] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0dd094f1-4e17-4137-b0a9-4224c597f8a7 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Lock "79c4d6e8-9999-4b0c-98d3-bbfd4d61e660" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 950.607021] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0dd094f1-4e17-4137-b0a9-4224c597f8a7 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Acquiring lock "79c4d6e8-9999-4b0c-98d3-bbfd4d61e660-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 950.607216] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0dd094f1-4e17-4137-b0a9-4224c597f8a7 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Lock "79c4d6e8-9999-4b0c-98d3-bbfd4d61e660-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 950.607548] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0dd094f1-4e17-4137-b0a9-4224c597f8a7 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Lock "79c4d6e8-9999-4b0c-98d3-bbfd4d61e660-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.609687] env[63088]: INFO nova.compute.manager [None req-0dd094f1-4e17-4137-b0a9-4224c597f8a7 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 79c4d6e8-9999-4b0c-98d3-bbfd4d61e660] Terminating instance [ 950.611492] env[63088]: DEBUG nova.compute.manager [None req-0dd094f1-4e17-4137-b0a9-4224c597f8a7 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 79c4d6e8-9999-4b0c-98d3-bbfd4d61e660] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 950.611877] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-0dd094f1-4e17-4137-b0a9-4224c597f8a7 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 79c4d6e8-9999-4b0c-98d3-bbfd4d61e660] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 950.612500] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce2305d0-5816-4070-a81e-30ed3494244e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.623312] env[63088]: DEBUG oslo_vmware.api [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285261, 'name': PowerOnVM_Task, 'duration_secs': 0.817879} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.625362] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 7cd45e52-c6be-4ec6-9710-fa99e29cf872] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 950.625575] env[63088]: INFO nova.compute.manager [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 7cd45e52-c6be-4ec6-9710-fa99e29cf872] Took 11.18 seconds to spawn the instance on the hypervisor. [ 950.625761] env[63088]: DEBUG nova.compute.manager [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 7cd45e52-c6be-4ec6-9710-fa99e29cf872] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 950.626065] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-0dd094f1-4e17-4137-b0a9-4224c597f8a7 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 79c4d6e8-9999-4b0c-98d3-bbfd4d61e660] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 950.627035] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-464f37ea-1a50-4ff2-9129-9372de0b6a79 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.629368] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-99f450bc-5bcd-4982-962e-21ced5c8e1b2 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.633016] env[63088]: INFO nova.compute.claims [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2ad33254-8030-4454-b023-3359071f85de] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 950.645097] env[63088]: DEBUG oslo_vmware.api [None req-0dd094f1-4e17-4137-b0a9-4224c597f8a7 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Waiting for the task: (returnval){ [ 950.645097] env[63088]: value = "task-1285264" [ 950.645097] env[63088]: _type = "Task" [ 950.645097] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.653491] env[63088]: DEBUG oslo_vmware.api [None req-0dd094f1-4e17-4137-b0a9-4224c597f8a7 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285264, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.675492] env[63088]: DEBUG oslo_vmware.api [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52e89bd6-cc5c-dd91-047a-ff4ea55f0cf7, 'name': SearchDatastore_Task, 'duration_secs': 0.016605} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.676323] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-588436f6-1e96-41c8-bcaf-2738ce940bcb {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.682180] env[63088]: DEBUG oslo_vmware.api [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Waiting for the task: (returnval){ [ 950.682180] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]526bc274-1d85-e3af-d66e-c298ec98c820" [ 950.682180] env[63088]: _type = "Task" [ 950.682180] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.690937] env[63088]: DEBUG oslo_vmware.api [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]526bc274-1d85-e3af-d66e-c298ec98c820, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.708840] env[63088]: DEBUG oslo_vmware.api [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Task: {'id': task-1285263, 'name': Rename_Task, 'duration_secs': 0.234652} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.709103] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] [instance: 8205f9e1-effd-442d-990b-bf5a2a9e6cdd] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 950.709348] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e979b62b-be57-4dc0-84cd-9fc7051e5c05 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.715901] env[63088]: DEBUG oslo_vmware.api [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Waiting for the task: (returnval){ [ 950.715901] env[63088]: value = "task-1285265" [ 950.715901] env[63088]: _type = "Task" [ 950.715901] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.728543] env[63088]: DEBUG oslo_vmware.api [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Task: {'id': task-1285265, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.731653] env[63088]: DEBUG oslo_vmware.api [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285262, 'name': CloneVM_Task} progress is 94%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.752961] env[63088]: DEBUG nova.compute.manager [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] [instance: 4a56d74c-cb1f-4edf-99f0-4b54bf357a3a] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 950.782316] env[63088]: DEBUG nova.virt.hardware [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 950.782719] env[63088]: DEBUG nova.virt.hardware [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 950.782979] env[63088]: DEBUG nova.virt.hardware [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 950.783216] env[63088]: DEBUG nova.virt.hardware [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 950.783388] env[63088]: DEBUG nova.virt.hardware [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 950.783550] env[63088]: DEBUG nova.virt.hardware [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 950.783791] env[63088]: DEBUG nova.virt.hardware [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 950.783962] env[63088]: DEBUG nova.virt.hardware [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 950.784154] env[63088]: DEBUG nova.virt.hardware [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 950.784341] env[63088]: DEBUG nova.virt.hardware [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 950.784527] env[63088]: DEBUG nova.virt.hardware [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 950.785451] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7598682-fe9f-4f13-a470-e82627a33484 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.795824] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b25e0d2-616b-42c6-896c-664b143766fd {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.809672] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] [instance: 4a56d74c-cb1f-4edf-99f0-4b54bf357a3a] Instance VIF info [] {{(pid=63088) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 950.816180] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Creating folder: Project (9fae6350d5ca41d1a3c7700fbf8d1ca5). Parent ref: group-v275816. {{(pid=63088) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 950.816902] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4aca34f8-6c8b-4354-9642-d396f1c6b23d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.826480] env[63088]: INFO nova.virt.vmwareapi.vm_util [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Created folder: Project (9fae6350d5ca41d1a3c7700fbf8d1ca5) in parent group-v275816. [ 950.826687] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Creating folder: Instances. Parent ref: group-v275923. {{(pid=63088) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 950.826927] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-90ccb917-7e8b-4c01-86ce-94d07f314e3a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.834916] env[63088]: INFO nova.virt.vmwareapi.vm_util [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Created folder: Instances in parent group-v275923. [ 950.835166] env[63088]: DEBUG oslo.service.loopingcall [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 950.835369] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4a56d74c-cb1f-4edf-99f0-4b54bf357a3a] Creating VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 950.835572] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3058f0a8-20e3-4dbf-85e0-a15c91c78021 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.851384] env[63088]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 950.851384] env[63088]: value = "task-1285268" [ 950.851384] env[63088]: _type = "Task" [ 950.851384] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.858492] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285268, 'name': CreateVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.067757] env[63088]: DEBUG oslo_concurrency.lockutils [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Releasing lock "refresh_cache-49ee4180-6322-4003-a5b1-f2a91d190290" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 951.068422] env[63088]: DEBUG nova.compute.manager [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] [instance: 49ee4180-6322-4003-a5b1-f2a91d190290] Instance network_info: |[{"id": "15b6a284-5552-4594-97bf-23b7ec018f5d", "address": "fa:16:3e:2f:58:37", "network": {"id": "b16c35d8-18ac-44b6-9aed-4e3a48d8b6d0", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-1225586833-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ef5fcb02afd04a42acbc92a0b89ea294", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "42f08482-a1da-405d-9918-d733d9f5173c", "external-id": "nsx-vlan-transportzone-381", "segmentation_id": 381, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap15b6a284-55", "ovs_interfaceid": "15b6a284-5552-4594-97bf-23b7ec018f5d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 951.068891] env[63088]: DEBUG oslo_concurrency.lockutils [req-ee1c8620-275a-4e5c-b0d8-ba3750eb22dc req-c32e500a-e53a-4f65-8240-78841d878bb4 service nova] Acquired lock "refresh_cache-49ee4180-6322-4003-a5b1-f2a91d190290" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 951.068891] env[63088]: DEBUG nova.network.neutron [req-ee1c8620-275a-4e5c-b0d8-ba3750eb22dc req-c32e500a-e53a-4f65-8240-78841d878bb4 service nova] [instance: 49ee4180-6322-4003-a5b1-f2a91d190290] Refreshing network info cache for port 15b6a284-5552-4594-97bf-23b7ec018f5d {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 951.070476] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] [instance: 49ee4180-6322-4003-a5b1-f2a91d190290] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2f:58:37', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '42f08482-a1da-405d-9918-d733d9f5173c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '15b6a284-5552-4594-97bf-23b7ec018f5d', 'vif_model': 'vmxnet3'}] {{(pid=63088) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 951.080276] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Creating folder: Project (ef5fcb02afd04a42acbc92a0b89ea294). Parent ref: group-v275816. {{(pid=63088) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 951.080771] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-079542ef-3706-439a-a155-1ff7a399a457 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.093733] env[63088]: INFO nova.virt.vmwareapi.vm_util [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Created folder: Project (ef5fcb02afd04a42acbc92a0b89ea294) in parent group-v275816. [ 951.094092] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Creating folder: Instances. Parent ref: group-v275926. {{(pid=63088) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 951.094447] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3c8009cc-094d-4832-960a-5f115b4e2efe {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.104018] env[63088]: INFO nova.virt.vmwareapi.vm_util [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Created folder: Instances in parent group-v275926. [ 951.104302] env[63088]: DEBUG oslo.service.loopingcall [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 951.104508] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 49ee4180-6322-4003-a5b1-f2a91d190290] Creating VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 951.104728] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6eca641e-cad6-4340-a294-fd9727f87447 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.124886] env[63088]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 951.124886] env[63088]: value = "task-1285271" [ 951.124886] env[63088]: _type = "Task" [ 951.124886] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.133031] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285271, 'name': CreateVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.138860] env[63088]: INFO nova.compute.resource_tracker [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2ad33254-8030-4454-b023-3359071f85de] Updating resource usage from migration 391d39ab-da7f-476e-bf0e-659212e03687 [ 951.157842] env[63088]: INFO nova.compute.manager [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 7cd45e52-c6be-4ec6-9710-fa99e29cf872] Took 23.22 seconds to build instance. [ 951.167037] env[63088]: DEBUG oslo_vmware.api [None req-0dd094f1-4e17-4137-b0a9-4224c597f8a7 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285264, 'name': PowerOffVM_Task, 'duration_secs': 0.188794} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.167037] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-0dd094f1-4e17-4137-b0a9-4224c597f8a7 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 79c4d6e8-9999-4b0c-98d3-bbfd4d61e660] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 951.167449] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-0dd094f1-4e17-4137-b0a9-4224c597f8a7 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 79c4d6e8-9999-4b0c-98d3-bbfd4d61e660] Unregistering the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 951.167449] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-88e7e56b-db45-4e08-b521-dd58d378e79d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.192774] env[63088]: DEBUG oslo_vmware.api [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]526bc274-1d85-e3af-d66e-c298ec98c820, 'name': SearchDatastore_Task, 'duration_secs': 0.009242} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.195435] env[63088]: DEBUG oslo_concurrency.lockutils [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 951.195735] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore2] d6fef3ef-fec8-4929-b9b8-5e63306aa51d/d6fef3ef-fec8-4929-b9b8-5e63306aa51d.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 951.196193] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-485f0fbe-3195-4846-b560-463ebeb9e1c3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.202697] env[63088]: DEBUG oslo_vmware.api [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Waiting for the task: (returnval){ [ 951.202697] env[63088]: value = "task-1285273" [ 951.202697] env[63088]: _type = "Task" [ 951.202697] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.215449] env[63088]: DEBUG oslo_vmware.api [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285273, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.231468] env[63088]: DEBUG oslo_vmware.api [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Task: {'id': task-1285265, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.231849] env[63088]: DEBUG oslo_vmware.api [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285262, 'name': CloneVM_Task} progress is 95%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.287968] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-0dd094f1-4e17-4137-b0a9-4224c597f8a7 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 79c4d6e8-9999-4b0c-98d3-bbfd4d61e660] Unregistered the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 951.288248] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-0dd094f1-4e17-4137-b0a9-4224c597f8a7 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 79c4d6e8-9999-4b0c-98d3-bbfd4d61e660] Deleting contents of the VM from datastore datastore2 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 951.288441] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-0dd094f1-4e17-4137-b0a9-4224c597f8a7 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Deleting the datastore file [datastore2] 79c4d6e8-9999-4b0c-98d3-bbfd4d61e660 {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 951.288722] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-edc33fe5-3809-4266-ae85-57a41b067e9e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.296445] env[63088]: DEBUG oslo_vmware.api [None req-0dd094f1-4e17-4137-b0a9-4224c597f8a7 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Waiting for the task: (returnval){ [ 951.296445] env[63088]: value = "task-1285274" [ 951.296445] env[63088]: _type = "Task" [ 951.296445] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.305980] env[63088]: DEBUG oslo_vmware.api [None req-0dd094f1-4e17-4137-b0a9-4224c597f8a7 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285274, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.344850] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90c71352-dfa5-4f04-ae49-9ab086853fc4 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.356381] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ba6c2f4-70de-4565-9da1-b86c94b1fbf2 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.364764] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285268, 'name': CreateVM_Task, 'duration_secs': 0.355964} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.390927] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4a56d74c-cb1f-4edf-99f0-4b54bf357a3a] Created VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 951.391871] env[63088]: DEBUG oslo_concurrency.lockutils [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 951.392101] env[63088]: DEBUG oslo_concurrency.lockutils [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 951.392481] env[63088]: DEBUG oslo_concurrency.lockutils [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 951.393673] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5a1adb7-2ac3-4b3a-9f14-2959c2d70dd2 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.396935] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c3060bed-7f80-4f0a-9cd6-939dc0940869 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.405608] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4305240e-1cca-430c-ae0c-64ef1b1298c0 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.410375] env[63088]: DEBUG oslo_vmware.api [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Waiting for the task: (returnval){ [ 951.410375] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52ad5a6a-d3fb-6870-1343-cdcbea6c9273" [ 951.410375] env[63088]: _type = "Task" [ 951.410375] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.423715] env[63088]: DEBUG nova.compute.provider_tree [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 951.430462] env[63088]: DEBUG oslo_vmware.api [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52ad5a6a-d3fb-6870-1343-cdcbea6c9273, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.432837] env[63088]: DEBUG nova.compute.manager [req-e526d7d4-4109-4255-ad34-9a08b1883a97 req-5dea342c-583c-429e-a00d-a3c670de99a1 service nova] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] Received event network-changed-88203cff-5be6-4261-a168-197e6be13cc3 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 951.433063] env[63088]: DEBUG nova.compute.manager [req-e526d7d4-4109-4255-ad34-9a08b1883a97 req-5dea342c-583c-429e-a00d-a3c670de99a1 service nova] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] Refreshing instance network info cache due to event network-changed-88203cff-5be6-4261-a168-197e6be13cc3. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 951.433291] env[63088]: DEBUG oslo_concurrency.lockutils [req-e526d7d4-4109-4255-ad34-9a08b1883a97 req-5dea342c-583c-429e-a00d-a3c670de99a1 service nova] Acquiring lock "refresh_cache-72c07d1e-cbb6-4875-9b7f-2649aa25eee5" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 951.433446] env[63088]: DEBUG oslo_concurrency.lockutils [req-e526d7d4-4109-4255-ad34-9a08b1883a97 req-5dea342c-583c-429e-a00d-a3c670de99a1 service nova] Acquired lock "refresh_cache-72c07d1e-cbb6-4875-9b7f-2649aa25eee5" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 951.433638] env[63088]: DEBUG nova.network.neutron [req-e526d7d4-4109-4255-ad34-9a08b1883a97 req-5dea342c-583c-429e-a00d-a3c670de99a1 service nova] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] Refreshing network info cache for port 88203cff-5be6-4261-a168-197e6be13cc3 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 951.640172] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285271, 'name': CreateVM_Task, 'duration_secs': 0.366922} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.640470] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 49ee4180-6322-4003-a5b1-f2a91d190290] Created VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 951.641190] env[63088]: DEBUG oslo_concurrency.lockutils [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 951.662732] env[63088]: DEBUG oslo_concurrency.lockutils [None req-130fd53d-4486-4500-8f5b-3fee582b9ff0 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lock "7cd45e52-c6be-4ec6-9710-fa99e29cf872" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.733s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.663175] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b3a20ade-2970-49c5-b2b9-f8e2a795f60b tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquiring lock "72c07d1e-cbb6-4875-9b7f-2649aa25eee5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 951.712380] env[63088]: DEBUG oslo_vmware.api [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285273, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.482981} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.712644] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore2] d6fef3ef-fec8-4929-b9b8-5e63306aa51d/d6fef3ef-fec8-4929-b9b8-5e63306aa51d.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 951.712859] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Extending root virtual disk to 1048576 {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 951.713151] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-aa2e84b7-f806-41f2-a662-5938e36734f7 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.722259] env[63088]: DEBUG oslo_vmware.api [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Waiting for the task: (returnval){ [ 951.722259] env[63088]: value = "task-1285275" [ 951.722259] env[63088]: _type = "Task" [ 951.722259] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.734078] env[63088]: DEBUG oslo_vmware.api [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Task: {'id': task-1285265, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.738726] env[63088]: DEBUG oslo_vmware.api [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285262, 'name': CloneVM_Task, 'duration_secs': 1.680351} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.741453] env[63088]: INFO nova.virt.vmwareapi.vmops [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Created linked-clone VM from snapshot [ 951.741730] env[63088]: DEBUG oslo_vmware.api [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285275, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.744436] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8aac1f7-76cc-48df-9ef6-4b0e80775ff7 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.752124] env[63088]: DEBUG nova.virt.vmwareapi.images [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Uploading image e3652b5e-39c5-4fe1-8b63-944e4527acd9 {{(pid=63088) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 951.775990] env[63088]: DEBUG oslo_vmware.rw_handles [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 951.775990] env[63088]: value = "vm-275922" [ 951.775990] env[63088]: _type = "VirtualMachine" [ 951.775990] env[63088]: }. {{(pid=63088) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 951.776340] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-80ab2c54-edcd-46a8-95a1-7eb8ab9404a7 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.783628] env[63088]: DEBUG oslo_vmware.rw_handles [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Lease: (returnval){ [ 951.783628] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52baf4db-3f92-ea12-355f-4a7944273f56" [ 951.783628] env[63088]: _type = "HttpNfcLease" [ 951.783628] env[63088]: } obtained for exporting VM: (result){ [ 951.783628] env[63088]: value = "vm-275922" [ 951.783628] env[63088]: _type = "VirtualMachine" [ 951.783628] env[63088]: }. {{(pid=63088) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 951.783945] env[63088]: DEBUG oslo_vmware.api [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Waiting for the lease: (returnval){ [ 951.783945] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52baf4db-3f92-ea12-355f-4a7944273f56" [ 951.783945] env[63088]: _type = "HttpNfcLease" [ 951.783945] env[63088]: } to be ready. {{(pid=63088) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 951.789679] env[63088]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 951.789679] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52baf4db-3f92-ea12-355f-4a7944273f56" [ 951.789679] env[63088]: _type = "HttpNfcLease" [ 951.789679] env[63088]: } is initializing. {{(pid=63088) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 951.804767] env[63088]: DEBUG oslo_vmware.api [None req-0dd094f1-4e17-4137-b0a9-4224c597f8a7 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285274, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.430137} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.804993] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-0dd094f1-4e17-4137-b0a9-4224c597f8a7 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 951.805208] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-0dd094f1-4e17-4137-b0a9-4224c597f8a7 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 79c4d6e8-9999-4b0c-98d3-bbfd4d61e660] Deleted contents of the VM from datastore datastore2 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 951.805383] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-0dd094f1-4e17-4137-b0a9-4224c597f8a7 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 79c4d6e8-9999-4b0c-98d3-bbfd4d61e660] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 951.805556] env[63088]: INFO nova.compute.manager [None req-0dd094f1-4e17-4137-b0a9-4224c597f8a7 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 79c4d6e8-9999-4b0c-98d3-bbfd4d61e660] Took 1.19 seconds to destroy the instance on the hypervisor. [ 951.806033] env[63088]: DEBUG oslo.service.loopingcall [None req-0dd094f1-4e17-4137-b0a9-4224c597f8a7 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 951.806033] env[63088]: DEBUG nova.compute.manager [-] [instance: 79c4d6e8-9999-4b0c-98d3-bbfd4d61e660] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 951.806199] env[63088]: DEBUG nova.network.neutron [-] [instance: 79c4d6e8-9999-4b0c-98d3-bbfd4d61e660] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 951.923555] env[63088]: DEBUG oslo_vmware.api [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52ad5a6a-d3fb-6870-1343-cdcbea6c9273, 'name': SearchDatastore_Task, 'duration_secs': 0.026684} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.924132] env[63088]: DEBUG oslo_concurrency.lockutils [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 951.924698] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] [instance: 4a56d74c-cb1f-4edf-99f0-4b54bf357a3a] Processing image 1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 951.925038] env[63088]: DEBUG oslo_concurrency.lockutils [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 951.925308] env[63088]: DEBUG oslo_concurrency.lockutils [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 951.925866] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 951.926441] env[63088]: DEBUG oslo_concurrency.lockutils [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 951.927865] env[63088]: DEBUG oslo_concurrency.lockutils [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 951.930473] env[63088]: DEBUG nova.scheduler.client.report [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 951.937512] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-58fb5c4a-b2b2-4ff2-967a-0352d71b0082 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.941318] env[63088]: DEBUG nova.network.neutron [req-ee1c8620-275a-4e5c-b0d8-ba3750eb22dc req-c32e500a-e53a-4f65-8240-78841d878bb4 service nova] [instance: 49ee4180-6322-4003-a5b1-f2a91d190290] Updated VIF entry in instance network info cache for port 15b6a284-5552-4594-97bf-23b7ec018f5d. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 951.941715] env[63088]: DEBUG nova.network.neutron [req-ee1c8620-275a-4e5c-b0d8-ba3750eb22dc req-c32e500a-e53a-4f65-8240-78841d878bb4 service nova] [instance: 49ee4180-6322-4003-a5b1-f2a91d190290] Updating instance_info_cache with network_info: [{"id": "15b6a284-5552-4594-97bf-23b7ec018f5d", "address": "fa:16:3e:2f:58:37", "network": {"id": "b16c35d8-18ac-44b6-9aed-4e3a48d8b6d0", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-1225586833-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ef5fcb02afd04a42acbc92a0b89ea294", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "42f08482-a1da-405d-9918-d733d9f5173c", "external-id": "nsx-vlan-transportzone-381", "segmentation_id": 381, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap15b6a284-55", "ovs_interfaceid": "15b6a284-5552-4594-97bf-23b7ec018f5d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 951.943085] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-412567f7-2b83-4773-8338-8ca7aad9d14a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.950799] env[63088]: DEBUG oslo_vmware.api [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Waiting for the task: (returnval){ [ 951.950799] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52f77a85-3eb1-a405-5070-bad55f6ad9e7" [ 951.950799] env[63088]: _type = "Task" [ 951.950799] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.954856] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 951.955041] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 951.956090] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fb2c3701-47e8-4cb8-9977-ecf6a22c677c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.962474] env[63088]: DEBUG oslo_vmware.api [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52f77a85-3eb1-a405-5070-bad55f6ad9e7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.965457] env[63088]: DEBUG oslo_vmware.api [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Waiting for the task: (returnval){ [ 951.965457] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52fbfa14-3d0a-fb3b-dbe3-e9e8aa418c6f" [ 951.965457] env[63088]: _type = "Task" [ 951.965457] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.974735] env[63088]: DEBUG oslo_vmware.api [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52fbfa14-3d0a-fb3b-dbe3-e9e8aa418c6f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.114262] env[63088]: DEBUG oslo_concurrency.lockutils [None req-2edfcd9b-169c-4444-b2ca-db2a6c7d48b5 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquiring lock "7cd45e52-c6be-4ec6-9710-fa99e29cf872" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.114566] env[63088]: DEBUG oslo_concurrency.lockutils [None req-2edfcd9b-169c-4444-b2ca-db2a6c7d48b5 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lock "7cd45e52-c6be-4ec6-9710-fa99e29cf872" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.114822] env[63088]: DEBUG oslo_concurrency.lockutils [None req-2edfcd9b-169c-4444-b2ca-db2a6c7d48b5 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquiring lock "7cd45e52-c6be-4ec6-9710-fa99e29cf872-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.115046] env[63088]: DEBUG oslo_concurrency.lockutils [None req-2edfcd9b-169c-4444-b2ca-db2a6c7d48b5 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lock "7cd45e52-c6be-4ec6-9710-fa99e29cf872-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.115216] env[63088]: DEBUG oslo_concurrency.lockutils [None req-2edfcd9b-169c-4444-b2ca-db2a6c7d48b5 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lock "7cd45e52-c6be-4ec6-9710-fa99e29cf872-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 952.121630] env[63088]: INFO nova.compute.manager [None req-2edfcd9b-169c-4444-b2ca-db2a6c7d48b5 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 7cd45e52-c6be-4ec6-9710-fa99e29cf872] Terminating instance [ 952.123481] env[63088]: DEBUG nova.compute.manager [None req-2edfcd9b-169c-4444-b2ca-db2a6c7d48b5 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 7cd45e52-c6be-4ec6-9710-fa99e29cf872] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 952.123685] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-2edfcd9b-169c-4444-b2ca-db2a6c7d48b5 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 7cd45e52-c6be-4ec6-9710-fa99e29cf872] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 952.124543] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b10756e9-bc17-4cef-bcee-94cc6e20c85f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.135629] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-2edfcd9b-169c-4444-b2ca-db2a6c7d48b5 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 7cd45e52-c6be-4ec6-9710-fa99e29cf872] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 952.135629] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4407cf0d-3951-4c2a-952a-0165bfa161cb {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.141473] env[63088]: DEBUG oslo_vmware.api [None req-2edfcd9b-169c-4444-b2ca-db2a6c7d48b5 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Waiting for the task: (returnval){ [ 952.141473] env[63088]: value = "task-1285277" [ 952.141473] env[63088]: _type = "Task" [ 952.141473] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.151412] env[63088]: DEBUG oslo_vmware.api [None req-2edfcd9b-169c-4444-b2ca-db2a6c7d48b5 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285277, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.169674] env[63088]: DEBUG nova.compute.manager [req-d07d456c-6a15-441e-b79d-5c8d54a7c92c req-7ef1f038-fda3-41eb-ad42-983fbcc9d9e5 service nova] [instance: 79c4d6e8-9999-4b0c-98d3-bbfd4d61e660] Received event network-vif-deleted-933dd0ba-3083-4fec-bc5f-1270e8761e92 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 952.170065] env[63088]: INFO nova.compute.manager [req-d07d456c-6a15-441e-b79d-5c8d54a7c92c req-7ef1f038-fda3-41eb-ad42-983fbcc9d9e5 service nova] [instance: 79c4d6e8-9999-4b0c-98d3-bbfd4d61e660] Neutron deleted interface 933dd0ba-3083-4fec-bc5f-1270e8761e92; detaching it from the instance and deleting it from the info cache [ 952.170437] env[63088]: DEBUG nova.network.neutron [req-d07d456c-6a15-441e-b79d-5c8d54a7c92c req-7ef1f038-fda3-41eb-ad42-983fbcc9d9e5 service nova] [instance: 79c4d6e8-9999-4b0c-98d3-bbfd4d61e660] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 952.231999] env[63088]: DEBUG oslo_vmware.api [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Task: {'id': task-1285265, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.236862] env[63088]: DEBUG oslo_vmware.api [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285275, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065836} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.241019] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Extended root virtual disk {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 952.241019] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89a4e47c-f23c-4aaf-8a0a-6934bbdc397c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.242264] env[63088]: DEBUG nova.network.neutron [req-e526d7d4-4109-4255-ad34-9a08b1883a97 req-5dea342c-583c-429e-a00d-a3c670de99a1 service nova] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] Updated VIF entry in instance network info cache for port 88203cff-5be6-4261-a168-197e6be13cc3. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 952.242732] env[63088]: DEBUG nova.network.neutron [req-e526d7d4-4109-4255-ad34-9a08b1883a97 req-5dea342c-583c-429e-a00d-a3c670de99a1 service nova] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] Updating instance_info_cache with network_info: [{"id": "88203cff-5be6-4261-a168-197e6be13cc3", "address": "fa:16:3e:e8:83:1f", "network": {"id": "5b801381-6bb3-45cd-b81f-92b5d757e080", "bridge": null, "label": "tempest-DeleteServersTestJSON-1481457494-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e48f62dee9ad4e9b94b67bd871db5c63", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap88203cff-5b", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 952.267931] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Reconfiguring VM instance instance-00000058 to attach disk [datastore2] d6fef3ef-fec8-4929-b9b8-5e63306aa51d/d6fef3ef-fec8-4929-b9b8-5e63306aa51d.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 952.269035] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6da55180-a988-488a-8454-7e17d44ef13c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.289980] env[63088]: DEBUG oslo_vmware.api [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Waiting for the task: (returnval){ [ 952.289980] env[63088]: value = "task-1285278" [ 952.289980] env[63088]: _type = "Task" [ 952.289980] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.292399] env[63088]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 952.292399] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52baf4db-3f92-ea12-355f-4a7944273f56" [ 952.292399] env[63088]: _type = "HttpNfcLease" [ 952.292399] env[63088]: } is ready. {{(pid=63088) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 952.295295] env[63088]: DEBUG oslo_vmware.rw_handles [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 952.295295] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52baf4db-3f92-ea12-355f-4a7944273f56" [ 952.295295] env[63088]: _type = "HttpNfcLease" [ 952.295295] env[63088]: }. {{(pid=63088) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 952.296009] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a612e13e-24b8-4e23-85ab-866e16449e9e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.304983] env[63088]: DEBUG oslo_vmware.rw_handles [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/521c4025-547d-e28f-7ab0-4f22d06e84bc/disk-0.vmdk from lease info. {{(pid=63088) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 952.305441] env[63088]: DEBUG oslo_vmware.rw_handles [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/521c4025-547d-e28f-7ab0-4f22d06e84bc/disk-0.vmdk for reading. {{(pid=63088) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 952.309707] env[63088]: DEBUG oslo_vmware.api [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285278, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.406023] env[63088]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-9deb0133-c04f-462e-8a0e-0ab97668a23a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.437858] env[63088]: DEBUG oslo_concurrency.lockutils [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.313s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 952.438090] env[63088]: INFO nova.compute.manager [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2ad33254-8030-4454-b023-3359071f85de] Migrating [ 952.444295] env[63088]: DEBUG oslo_concurrency.lockutils [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.950s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.444539] env[63088]: DEBUG nova.objects.instance [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lazy-loading 'resources' on Instance uuid 72c07d1e-cbb6-4875-9b7f-2649aa25eee5 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 952.449613] env[63088]: DEBUG oslo_concurrency.lockutils [req-ee1c8620-275a-4e5c-b0d8-ba3750eb22dc req-c32e500a-e53a-4f65-8240-78841d878bb4 service nova] Releasing lock "refresh_cache-49ee4180-6322-4003-a5b1-f2a91d190290" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 952.470854] env[63088]: DEBUG oslo_vmware.api [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52f77a85-3eb1-a405-5070-bad55f6ad9e7, 'name': SearchDatastore_Task, 'duration_secs': 0.008799} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.474288] env[63088]: DEBUG oslo_concurrency.lockutils [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 952.474518] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] [instance: 49ee4180-6322-4003-a5b1-f2a91d190290] Processing image 1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 952.474727] env[63088]: DEBUG oslo_concurrency.lockutils [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 952.481643] env[63088]: DEBUG oslo_vmware.api [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52fbfa14-3d0a-fb3b-dbe3-e9e8aa418c6f, 'name': SearchDatastore_Task, 'duration_secs': 0.008134} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.483242] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9dddaf0e-812c-4b7b-96bc-e022d54903d2 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.488591] env[63088]: DEBUG oslo_vmware.api [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Waiting for the task: (returnval){ [ 952.488591] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]529efac3-da9d-91cc-015e-346c801f8263" [ 952.488591] env[63088]: _type = "Task" [ 952.488591] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.496213] env[63088]: DEBUG oslo_vmware.api [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]529efac3-da9d-91cc-015e-346c801f8263, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.648583] env[63088]: DEBUG nova.network.neutron [-] [instance: 79c4d6e8-9999-4b0c-98d3-bbfd4d61e660] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 952.655155] env[63088]: DEBUG oslo_vmware.api [None req-2edfcd9b-169c-4444-b2ca-db2a6c7d48b5 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285277, 'name': PowerOffVM_Task, 'duration_secs': 0.223308} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.655569] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-2edfcd9b-169c-4444-b2ca-db2a6c7d48b5 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 7cd45e52-c6be-4ec6-9710-fa99e29cf872] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 952.655678] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-2edfcd9b-169c-4444-b2ca-db2a6c7d48b5 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 7cd45e52-c6be-4ec6-9710-fa99e29cf872] Unregistering the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 952.655926] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c106d3cb-5271-4831-a591-801f304b4278 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.673359] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2f7e2279-bb91-4819-870d-1e4812c71af3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.683359] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bfc2dc6-c20a-451f-8678-75bad8b3406f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.714339] env[63088]: DEBUG nova.compute.manager [req-d07d456c-6a15-441e-b79d-5c8d54a7c92c req-7ef1f038-fda3-41eb-ad42-983fbcc9d9e5 service nova] [instance: 79c4d6e8-9999-4b0c-98d3-bbfd4d61e660] Detach interface failed, port_id=933dd0ba-3083-4fec-bc5f-1270e8761e92, reason: Instance 79c4d6e8-9999-4b0c-98d3-bbfd4d61e660 could not be found. {{(pid=63088) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 952.722535] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-2edfcd9b-169c-4444-b2ca-db2a6c7d48b5 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 7cd45e52-c6be-4ec6-9710-fa99e29cf872] Unregistered the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 952.722800] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-2edfcd9b-169c-4444-b2ca-db2a6c7d48b5 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 7cd45e52-c6be-4ec6-9710-fa99e29cf872] Deleting contents of the VM from datastore datastore2 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 952.723162] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-2edfcd9b-169c-4444-b2ca-db2a6c7d48b5 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Deleting the datastore file [datastore2] 7cd45e52-c6be-4ec6-9710-fa99e29cf872 {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 952.723833] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8b8ef62c-1a18-4f5d-a190-fdbab8837e55 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.730633] env[63088]: DEBUG oslo_vmware.api [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Task: {'id': task-1285265, 'name': PowerOnVM_Task} progress is 76%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.733445] env[63088]: DEBUG oslo_vmware.api [None req-2edfcd9b-169c-4444-b2ca-db2a6c7d48b5 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Waiting for the task: (returnval){ [ 952.733445] env[63088]: value = "task-1285280" [ 952.733445] env[63088]: _type = "Task" [ 952.733445] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.742281] env[63088]: DEBUG oslo_vmware.api [None req-2edfcd9b-169c-4444-b2ca-db2a6c7d48b5 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285280, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.747023] env[63088]: DEBUG oslo_concurrency.lockutils [req-e526d7d4-4109-4255-ad34-9a08b1883a97 req-5dea342c-583c-429e-a00d-a3c670de99a1 service nova] Releasing lock "refresh_cache-72c07d1e-cbb6-4875-9b7f-2649aa25eee5" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 952.803889] env[63088]: DEBUG oslo_vmware.api [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285278, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.947679] env[63088]: DEBUG nova.objects.instance [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lazy-loading 'numa_topology' on Instance uuid 72c07d1e-cbb6-4875-9b7f-2649aa25eee5 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 952.956084] env[63088]: DEBUG oslo_concurrency.lockutils [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquiring lock "refresh_cache-2ad33254-8030-4454-b023-3359071f85de" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 952.956302] env[63088]: DEBUG oslo_concurrency.lockutils [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquired lock "refresh_cache-2ad33254-8030-4454-b023-3359071f85de" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.956491] env[63088]: DEBUG nova.network.neutron [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2ad33254-8030-4454-b023-3359071f85de] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 952.999960] env[63088]: DEBUG oslo_vmware.api [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]529efac3-da9d-91cc-015e-346c801f8263, 'name': SearchDatastore_Task, 'duration_secs': 0.009155} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.000273] env[63088]: DEBUG oslo_concurrency.lockutils [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 953.000687] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] 4a56d74c-cb1f-4edf-99f0-4b54bf357a3a/4a56d74c-cb1f-4edf-99f0-4b54bf357a3a.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 953.001052] env[63088]: DEBUG oslo_concurrency.lockutils [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 953.001355] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 953.001867] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3cfb3e4a-d3ea-44a7-b436-a4bcf43f4f98 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.006154] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-38443b54-b299-4e9b-afaa-f534dc0968a5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.014254] env[63088]: DEBUG oslo_vmware.api [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Waiting for the task: (returnval){ [ 953.014254] env[63088]: value = "task-1285281" [ 953.014254] env[63088]: _type = "Task" [ 953.014254] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.015639] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 953.015911] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 953.019982] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8d8e6d7b-7e18-40ec-82b0-39f1ce2afb52 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.027603] env[63088]: DEBUG oslo_vmware.api [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Task: {'id': task-1285281, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.029657] env[63088]: DEBUG oslo_vmware.api [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Waiting for the task: (returnval){ [ 953.029657] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52929dc9-e457-3be6-0eb7-1dd1f2b9c2f1" [ 953.029657] env[63088]: _type = "Task" [ 953.029657] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.039565] env[63088]: DEBUG oslo_vmware.api [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52929dc9-e457-3be6-0eb7-1dd1f2b9c2f1, 'name': SearchDatastore_Task, 'duration_secs': 0.009798} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.040332] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-717d154a-3e39-4221-8c7e-8ca081bdffc9 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.045628] env[63088]: DEBUG oslo_vmware.api [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Waiting for the task: (returnval){ [ 953.045628] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52e1e1c6-4b7d-9459-ba56-c77d6ef2af68" [ 953.045628] env[63088]: _type = "Task" [ 953.045628] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.053467] env[63088]: DEBUG oslo_vmware.api [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52e1e1c6-4b7d-9459-ba56-c77d6ef2af68, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.150732] env[63088]: INFO nova.compute.manager [-] [instance: 79c4d6e8-9999-4b0c-98d3-bbfd4d61e660] Took 1.34 seconds to deallocate network for instance. [ 953.230888] env[63088]: DEBUG oslo_vmware.api [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Task: {'id': task-1285265, 'name': PowerOnVM_Task, 'duration_secs': 2.260665} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.231276] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] [instance: 8205f9e1-effd-442d-990b-bf5a2a9e6cdd] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 953.232120] env[63088]: INFO nova.compute.manager [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] [instance: 8205f9e1-effd-442d-990b-bf5a2a9e6cdd] Took 9.64 seconds to spawn the instance on the hypervisor. [ 953.232543] env[63088]: DEBUG nova.compute.manager [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] [instance: 8205f9e1-effd-442d-990b-bf5a2a9e6cdd] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 953.233475] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea38caf1-5351-46b9-907e-9354b451b809 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.249474] env[63088]: DEBUG oslo_vmware.api [None req-2edfcd9b-169c-4444-b2ca-db2a6c7d48b5 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285280, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.138121} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.250085] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-2edfcd9b-169c-4444-b2ca-db2a6c7d48b5 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 953.250541] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-2edfcd9b-169c-4444-b2ca-db2a6c7d48b5 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 7cd45e52-c6be-4ec6-9710-fa99e29cf872] Deleted contents of the VM from datastore datastore2 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 953.250770] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-2edfcd9b-169c-4444-b2ca-db2a6c7d48b5 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 7cd45e52-c6be-4ec6-9710-fa99e29cf872] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 953.251067] env[63088]: INFO nova.compute.manager [None req-2edfcd9b-169c-4444-b2ca-db2a6c7d48b5 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: 7cd45e52-c6be-4ec6-9710-fa99e29cf872] Took 1.13 seconds to destroy the instance on the hypervisor. [ 953.251392] env[63088]: DEBUG oslo.service.loopingcall [None req-2edfcd9b-169c-4444-b2ca-db2a6c7d48b5 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 953.252487] env[63088]: DEBUG nova.compute.manager [-] [instance: 7cd45e52-c6be-4ec6-9710-fa99e29cf872] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 953.252487] env[63088]: DEBUG nova.network.neutron [-] [instance: 7cd45e52-c6be-4ec6-9710-fa99e29cf872] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 953.304834] env[63088]: DEBUG oslo_vmware.api [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285278, 'name': ReconfigVM_Task, 'duration_secs': 0.617525} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.305957] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Reconfigured VM instance instance-00000058 to attach disk [datastore2] d6fef3ef-fec8-4929-b9b8-5e63306aa51d/d6fef3ef-fec8-4929-b9b8-5e63306aa51d.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 953.306104] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-250777b5-b16c-4fd4-b348-3aea717fc7e6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.313120] env[63088]: DEBUG oslo_vmware.api [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Waiting for the task: (returnval){ [ 953.313120] env[63088]: value = "task-1285282" [ 953.313120] env[63088]: _type = "Task" [ 953.313120] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.322463] env[63088]: DEBUG oslo_vmware.api [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285282, 'name': Rename_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.453734] env[63088]: DEBUG nova.objects.base [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Object Instance<72c07d1e-cbb6-4875-9b7f-2649aa25eee5> lazy-loaded attributes: resources,numa_topology {{(pid=63088) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 953.530540] env[63088]: DEBUG oslo_vmware.api [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Task: {'id': task-1285281, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.558564] env[63088]: DEBUG oslo_vmware.api [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52e1e1c6-4b7d-9459-ba56-c77d6ef2af68, 'name': SearchDatastore_Task, 'duration_secs': 0.008301} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.562326] env[63088]: DEBUG oslo_concurrency.lockutils [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 953.565778] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] 49ee4180-6322-4003-a5b1-f2a91d190290/49ee4180-6322-4003-a5b1-f2a91d190290.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 953.565778] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-13882dbf-8c7b-41a7-a6a7-39267b798e21 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.573018] env[63088]: DEBUG oslo_vmware.api [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Waiting for the task: (returnval){ [ 953.573018] env[63088]: value = "task-1285283" [ 953.573018] env[63088]: _type = "Task" [ 953.573018] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.580754] env[63088]: DEBUG oslo_vmware.api [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Task: {'id': task-1285283, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.657676] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0dd094f1-4e17-4137-b0a9-4224c597f8a7 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.729145] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9140e1a-54d7-42cd-b8a9-119d6db7cfe3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.739821] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-839bf58e-dc83-495a-9e80-4a277fcb38dd {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.783959] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6fcf873-7deb-420b-9c2f-6d8f2a0c6ab1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.787524] env[63088]: INFO nova.compute.manager [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] [instance: 8205f9e1-effd-442d-990b-bf5a2a9e6cdd] Took 19.61 seconds to build instance. [ 953.794859] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dfb87fa-8020-4721-a883-27cca96d725a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.810959] env[63088]: DEBUG nova.compute.provider_tree [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 953.824429] env[63088]: DEBUG oslo_vmware.api [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285282, 'name': Rename_Task} progress is 99%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.918877] env[63088]: DEBUG nova.network.neutron [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2ad33254-8030-4454-b023-3359071f85de] Updating instance_info_cache with network_info: [{"id": "1d2bee7e-bbc7-4b17-976b-47e7bbb268af", "address": "fa:16:3e:34:9a:ff", "network": {"id": "dff14a7f-0af1-4e4e-a498-86d7c9816e6b", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-762073671-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f33f2701fad94864a8c406a404bc0a42", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "21310d90-efbc-45a8-a97f-c4358606530f", "external-id": "nsx-vlan-transportzone-672", "segmentation_id": 672, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d2bee7e-bb", "ovs_interfaceid": "1d2bee7e-bbc7-4b17-976b-47e7bbb268af", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 954.026395] env[63088]: DEBUG oslo_vmware.api [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Task: {'id': task-1285281, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.571946} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.026851] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] 4a56d74c-cb1f-4edf-99f0-4b54bf357a3a/4a56d74c-cb1f-4edf-99f0-4b54bf357a3a.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 954.027083] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] [instance: 4a56d74c-cb1f-4edf-99f0-4b54bf357a3a] Extending root virtual disk to 1048576 {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 954.027550] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-56738851-2cfa-4d3e-9b76-08f6d5437417 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.034788] env[63088]: DEBUG oslo_vmware.api [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Waiting for the task: (returnval){ [ 954.034788] env[63088]: value = "task-1285284" [ 954.034788] env[63088]: _type = "Task" [ 954.034788] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.042731] env[63088]: DEBUG oslo_vmware.api [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Task: {'id': task-1285284, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.083514] env[63088]: DEBUG oslo_vmware.api [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Task: {'id': task-1285283, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.455444} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.084160] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] 49ee4180-6322-4003-a5b1-f2a91d190290/49ee4180-6322-4003-a5b1-f2a91d190290.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 954.084602] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] [instance: 49ee4180-6322-4003-a5b1-f2a91d190290] Extending root virtual disk to 1048576 {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 954.087226] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-495a37b3-e612-467b-8a8a-4cceb23e86ea {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.091806] env[63088]: DEBUG oslo_vmware.api [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Waiting for the task: (returnval){ [ 954.091806] env[63088]: value = "task-1285285" [ 954.091806] env[63088]: _type = "Task" [ 954.091806] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.100753] env[63088]: DEBUG oslo_vmware.api [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Task: {'id': task-1285285, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.208079] env[63088]: DEBUG nova.network.neutron [-] [instance: 7cd45e52-c6be-4ec6-9710-fa99e29cf872] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 954.265045] env[63088]: DEBUG nova.compute.manager [req-f688a910-98da-4ea8-8b56-d6c552431f0d req-2d3ba46a-0bf0-4788-8a1a-1b959322c303 service nova] [instance: 7cd45e52-c6be-4ec6-9710-fa99e29cf872] Received event network-vif-deleted-ba2a84ec-889d-4dc1-b6ca-95d59271b186 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 954.290035] env[63088]: DEBUG oslo_concurrency.lockutils [None req-e65faa2d-e673-41b3-ba64-70036bcc2388 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Lock "8205f9e1-effd-442d-990b-bf5a2a9e6cdd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.118s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.317028] env[63088]: DEBUG nova.scheduler.client.report [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 954.328359] env[63088]: DEBUG oslo_vmware.api [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285282, 'name': Rename_Task} progress is 99%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.422543] env[63088]: DEBUG oslo_concurrency.lockutils [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Releasing lock "refresh_cache-2ad33254-8030-4454-b023-3359071f85de" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 954.546194] env[63088]: DEBUG oslo_vmware.api [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Task: {'id': task-1285284, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.237154} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.546592] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] [instance: 4a56d74c-cb1f-4edf-99f0-4b54bf357a3a] Extended root virtual disk {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 954.547505] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2fab81c-9633-49eb-9726-5f3023f0b514 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.567973] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] [instance: 4a56d74c-cb1f-4edf-99f0-4b54bf357a3a] Reconfiguring VM instance instance-0000005a to attach disk [datastore1] 4a56d74c-cb1f-4edf-99f0-4b54bf357a3a/4a56d74c-cb1f-4edf-99f0-4b54bf357a3a.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 954.568744] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-388bf685-0fe7-4c39-98d9-e36023d0cab2 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.588897] env[63088]: DEBUG oslo_vmware.api [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Waiting for the task: (returnval){ [ 954.588897] env[63088]: value = "task-1285286" [ 954.588897] env[63088]: _type = "Task" [ 954.588897] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.598379] env[63088]: DEBUG oslo_vmware.api [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Task: {'id': task-1285286, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.603756] env[63088]: DEBUG oslo_vmware.api [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Task: {'id': task-1285285, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.710393] env[63088]: INFO nova.compute.manager [-] [instance: 7cd45e52-c6be-4ec6-9710-fa99e29cf872] Took 1.46 seconds to deallocate network for instance. [ 954.823954] env[63088]: DEBUG oslo_concurrency.lockutils [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.380s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.826711] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0dd094f1-4e17-4137-b0a9-4224c597f8a7 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.169s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.826952] env[63088]: DEBUG nova.objects.instance [None req-0dd094f1-4e17-4137-b0a9-4224c597f8a7 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Lazy-loading 'resources' on Instance uuid 79c4d6e8-9999-4b0c-98d3-bbfd4d61e660 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 954.833930] env[63088]: DEBUG oslo_vmware.api [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285282, 'name': Rename_Task, 'duration_secs': 1.187191} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.834222] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 954.835028] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6e755db3-285f-407d-b0f9-79b683e1a582 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.842764] env[63088]: DEBUG oslo_vmware.api [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Waiting for the task: (returnval){ [ 954.842764] env[63088]: value = "task-1285287" [ 954.842764] env[63088]: _type = "Task" [ 954.842764] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.846216] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d176fcb8-4997-4cd3-a198-d23e1a48be28 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Acquiring lock "8205f9e1-effd-442d-990b-bf5a2a9e6cdd" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.846499] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d176fcb8-4997-4cd3-a198-d23e1a48be28 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Lock "8205f9e1-effd-442d-990b-bf5a2a9e6cdd" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.846712] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d176fcb8-4997-4cd3-a198-d23e1a48be28 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Acquiring lock "8205f9e1-effd-442d-990b-bf5a2a9e6cdd-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.846898] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d176fcb8-4997-4cd3-a198-d23e1a48be28 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Lock "8205f9e1-effd-442d-990b-bf5a2a9e6cdd-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.847527] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d176fcb8-4997-4cd3-a198-d23e1a48be28 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Lock "8205f9e1-effd-442d-990b-bf5a2a9e6cdd-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.849469] env[63088]: INFO nova.compute.manager [None req-d176fcb8-4997-4cd3-a198-d23e1a48be28 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] [instance: 8205f9e1-effd-442d-990b-bf5a2a9e6cdd] Terminating instance [ 954.854828] env[63088]: DEBUG oslo_vmware.api [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285287, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.855202] env[63088]: DEBUG nova.compute.manager [None req-d176fcb8-4997-4cd3-a198-d23e1a48be28 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] [instance: 8205f9e1-effd-442d-990b-bf5a2a9e6cdd] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 954.855373] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-d176fcb8-4997-4cd3-a198-d23e1a48be28 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] [instance: 8205f9e1-effd-442d-990b-bf5a2a9e6cdd] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 954.856292] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4b04339-1e06-41d0-bfa6-ccd9442f9ffc {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.864019] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-d176fcb8-4997-4cd3-a198-d23e1a48be28 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] [instance: 8205f9e1-effd-442d-990b-bf5a2a9e6cdd] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 954.864925] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a2320514-ec93-4297-86cc-104d55b93c65 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.871068] env[63088]: DEBUG oslo_vmware.api [None req-d176fcb8-4997-4cd3-a198-d23e1a48be28 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Waiting for the task: (returnval){ [ 954.871068] env[63088]: value = "task-1285288" [ 954.871068] env[63088]: _type = "Task" [ 954.871068] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.879678] env[63088]: DEBUG oslo_vmware.api [None req-d176fcb8-4997-4cd3-a198-d23e1a48be28 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Task: {'id': task-1285288, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.103554] env[63088]: DEBUG oslo_vmware.api [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Task: {'id': task-1285286, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.106681] env[63088]: DEBUG oslo_vmware.api [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Task: {'id': task-1285285, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.770333} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.106981] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] [instance: 49ee4180-6322-4003-a5b1-f2a91d190290] Extended root virtual disk {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 955.107814] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02dd386e-f4ae-4368-be8d-6d115082cc43 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.131231] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] [instance: 49ee4180-6322-4003-a5b1-f2a91d190290] Reconfiguring VM instance instance-00000059 to attach disk [datastore1] 49ee4180-6322-4003-a5b1-f2a91d190290/49ee4180-6322-4003-a5b1-f2a91d190290.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 955.131619] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9545101f-650e-4d5d-8d3d-1cb3b61473c6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.152893] env[63088]: DEBUG oslo_vmware.api [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Waiting for the task: (returnval){ [ 955.152893] env[63088]: value = "task-1285289" [ 955.152893] env[63088]: _type = "Task" [ 955.152893] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.161946] env[63088]: DEBUG oslo_vmware.api [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Task: {'id': task-1285289, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.218361] env[63088]: DEBUG oslo_concurrency.lockutils [None req-2edfcd9b-169c-4444-b2ca-db2a6c7d48b5 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.336191] env[63088]: DEBUG oslo_concurrency.lockutils [None req-99f52060-06cf-419b-86b8-f818613cd389 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "72c07d1e-cbb6-4875-9b7f-2649aa25eee5" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 23.837s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.337129] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b3a20ade-2970-49c5-b2b9-f8e2a795f60b tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "72c07d1e-cbb6-4875-9b7f-2649aa25eee5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 3.674s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.337958] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b3a20ade-2970-49c5-b2b9-f8e2a795f60b tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquiring lock "72c07d1e-cbb6-4875-9b7f-2649aa25eee5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.337958] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b3a20ade-2970-49c5-b2b9-f8e2a795f60b tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "72c07d1e-cbb6-4875-9b7f-2649aa25eee5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.337958] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b3a20ade-2970-49c5-b2b9-f8e2a795f60b tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "72c07d1e-cbb6-4875-9b7f-2649aa25eee5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.340052] env[63088]: INFO nova.compute.manager [None req-b3a20ade-2970-49c5-b2b9-f8e2a795f60b tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] Terminating instance [ 955.341861] env[63088]: DEBUG nova.compute.manager [None req-b3a20ade-2970-49c5-b2b9-f8e2a795f60b tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 955.342118] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-b3a20ade-2970-49c5-b2b9-f8e2a795f60b tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 955.342662] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5e61319e-4abd-4617-b565-9ec30512544f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.355723] env[63088]: DEBUG oslo_vmware.api [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285287, 'name': PowerOnVM_Task} progress is 87%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.363112] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f10d0811-3867-44db-a8e7-54f34f238691 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.383653] env[63088]: DEBUG oslo_vmware.api [None req-d176fcb8-4997-4cd3-a198-d23e1a48be28 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Task: {'id': task-1285288, 'name': PowerOffVM_Task, 'duration_secs': 0.243189} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.383964] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-d176fcb8-4997-4cd3-a198-d23e1a48be28 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] [instance: 8205f9e1-effd-442d-990b-bf5a2a9e6cdd] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 955.384180] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-d176fcb8-4997-4cd3-a198-d23e1a48be28 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] [instance: 8205f9e1-effd-442d-990b-bf5a2a9e6cdd] Unregistering the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 955.384471] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-abc24b16-e958-4fd3-abec-09ec2935a627 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.399119] env[63088]: WARNING nova.virt.vmwareapi.vmops [None req-b3a20ade-2970-49c5-b2b9-f8e2a795f60b tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 72c07d1e-cbb6-4875-9b7f-2649aa25eee5 could not be found. [ 955.399119] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-b3a20ade-2970-49c5-b2b9-f8e2a795f60b tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 955.399119] env[63088]: INFO nova.compute.manager [None req-b3a20ade-2970-49c5-b2b9-f8e2a795f60b tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] Took 0.06 seconds to destroy the instance on the hypervisor. [ 955.399432] env[63088]: DEBUG oslo.service.loopingcall [None req-b3a20ade-2970-49c5-b2b9-f8e2a795f60b tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 955.399524] env[63088]: DEBUG nova.compute.manager [-] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 955.399620] env[63088]: DEBUG nova.network.neutron [-] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 955.498939] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-d176fcb8-4997-4cd3-a198-d23e1a48be28 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] [instance: 8205f9e1-effd-442d-990b-bf5a2a9e6cdd] Unregistered the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 955.499153] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-d176fcb8-4997-4cd3-a198-d23e1a48be28 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] [instance: 8205f9e1-effd-442d-990b-bf5a2a9e6cdd] Deleting contents of the VM from datastore datastore2 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 955.499347] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-d176fcb8-4997-4cd3-a198-d23e1a48be28 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Deleting the datastore file [datastore2] 8205f9e1-effd-442d-990b-bf5a2a9e6cdd {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 955.499618] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2e0e99c4-76c7-4396-9be2-7b08992a3882 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.508670] env[63088]: DEBUG oslo_vmware.api [None req-d176fcb8-4997-4cd3-a198-d23e1a48be28 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Waiting for the task: (returnval){ [ 955.508670] env[63088]: value = "task-1285291" [ 955.508670] env[63088]: _type = "Task" [ 955.508670] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.520070] env[63088]: DEBUG oslo_vmware.api [None req-d176fcb8-4997-4cd3-a198-d23e1a48be28 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Task: {'id': task-1285291, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.570962] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5cf2b54-0a2a-4774-b453-bdbef63c81fc {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.581713] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39029a73-b48e-4bbc-a339-9784042a30c6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.620436] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9edf1865-6244-4200-b134-5b02d9dd856e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.628544] env[63088]: DEBUG oslo_vmware.api [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Task: {'id': task-1285286, 'name': ReconfigVM_Task, 'duration_secs': 0.53554} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.630932] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] [instance: 4a56d74c-cb1f-4edf-99f0-4b54bf357a3a] Reconfigured VM instance instance-0000005a to attach disk [datastore1] 4a56d74c-cb1f-4edf-99f0-4b54bf357a3a/4a56d74c-cb1f-4edf-99f0-4b54bf357a3a.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 955.632035] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8a7139ba-a4f4-4518-be29-321d8c746ab8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.634450] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-701d4dd0-f70b-432d-b99a-f5088b84db58 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.650183] env[63088]: DEBUG nova.compute.provider_tree [None req-0dd094f1-4e17-4137-b0a9-4224c597f8a7 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 955.652795] env[63088]: DEBUG oslo_vmware.api [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Waiting for the task: (returnval){ [ 955.652795] env[63088]: value = "task-1285292" [ 955.652795] env[63088]: _type = "Task" [ 955.652795] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.665388] env[63088]: DEBUG oslo_vmware.api [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Task: {'id': task-1285292, 'name': Rename_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.671534] env[63088]: DEBUG oslo_vmware.api [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Task: {'id': task-1285289, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.852636] env[63088]: DEBUG oslo_vmware.api [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285287, 'name': PowerOnVM_Task, 'duration_secs': 0.748528} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.853046] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 955.853171] env[63088]: INFO nova.compute.manager [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Took 9.84 seconds to spawn the instance on the hypervisor. [ 955.853337] env[63088]: DEBUG nova.compute.manager [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 955.854144] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e673b49-4ad4-46b6-9193-7333c1baecb4 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.937640] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8e262ae-2fec-42a2-9275-979a837868e8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.960660] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2ad33254-8030-4454-b023-3359071f85de] Updating instance '2ad33254-8030-4454-b023-3359071f85de' progress to 0 {{(pid=63088) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 956.019478] env[63088]: DEBUG oslo_vmware.api [None req-d176fcb8-4997-4cd3-a198-d23e1a48be28 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Task: {'id': task-1285291, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.154891} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.019754] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-d176fcb8-4997-4cd3-a198-d23e1a48be28 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 956.019942] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-d176fcb8-4997-4cd3-a198-d23e1a48be28 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] [instance: 8205f9e1-effd-442d-990b-bf5a2a9e6cdd] Deleted contents of the VM from datastore datastore2 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 956.020190] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-d176fcb8-4997-4cd3-a198-d23e1a48be28 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] [instance: 8205f9e1-effd-442d-990b-bf5a2a9e6cdd] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 956.020345] env[63088]: INFO nova.compute.manager [None req-d176fcb8-4997-4cd3-a198-d23e1a48be28 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] [instance: 8205f9e1-effd-442d-990b-bf5a2a9e6cdd] Took 1.17 seconds to destroy the instance on the hypervisor. [ 956.020593] env[63088]: DEBUG oslo.service.loopingcall [None req-d176fcb8-4997-4cd3-a198-d23e1a48be28 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 956.020786] env[63088]: DEBUG nova.compute.manager [-] [instance: 8205f9e1-effd-442d-990b-bf5a2a9e6cdd] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 956.020881] env[63088]: DEBUG nova.network.neutron [-] [instance: 8205f9e1-effd-442d-990b-bf5a2a9e6cdd] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 956.159756] env[63088]: DEBUG nova.scheduler.client.report [None req-0dd094f1-4e17-4137-b0a9-4224c597f8a7 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 956.174244] env[63088]: DEBUG oslo_vmware.api [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Task: {'id': task-1285289, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.174873] env[63088]: DEBUG oslo_vmware.api [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Task: {'id': task-1285292, 'name': Rename_Task, 'duration_secs': 0.198627} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.175377] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] [instance: 4a56d74c-cb1f-4edf-99f0-4b54bf357a3a] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 956.175634] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6d7f7f4e-9325-4412-8791-28b5c2df71f4 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.183050] env[63088]: DEBUG oslo_vmware.api [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Waiting for the task: (returnval){ [ 956.183050] env[63088]: value = "task-1285293" [ 956.183050] env[63088]: _type = "Task" [ 956.183050] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.193201] env[63088]: DEBUG oslo_vmware.api [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Task: {'id': task-1285293, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.198953] env[63088]: DEBUG nova.network.neutron [-] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 956.298284] env[63088]: DEBUG nova.compute.manager [req-c69bd41f-a5ca-4962-bd0b-d886d680a967 req-a2e4eeb3-ab71-4e55-9623-48ad468e5aee service nova] [instance: 8205f9e1-effd-442d-990b-bf5a2a9e6cdd] Received event network-vif-deleted-faac8e4c-2334-4c13-ae85-13e55b44fbb6 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 956.298568] env[63088]: INFO nova.compute.manager [req-c69bd41f-a5ca-4962-bd0b-d886d680a967 req-a2e4eeb3-ab71-4e55-9623-48ad468e5aee service nova] [instance: 8205f9e1-effd-442d-990b-bf5a2a9e6cdd] Neutron deleted interface faac8e4c-2334-4c13-ae85-13e55b44fbb6; detaching it from the instance and deleting it from the info cache [ 956.298838] env[63088]: DEBUG nova.network.neutron [req-c69bd41f-a5ca-4962-bd0b-d886d680a967 req-a2e4eeb3-ab71-4e55-9623-48ad468e5aee service nova] [instance: 8205f9e1-effd-442d-990b-bf5a2a9e6cdd] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 956.374383] env[63088]: INFO nova.compute.manager [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Took 19.57 seconds to build instance. [ 956.467335] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2ad33254-8030-4454-b023-3359071f85de] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 956.467665] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1e8c3076-7ee0-4b8c-9e68-498b1f631f00 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.476414] env[63088]: DEBUG oslo_vmware.api [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Waiting for the task: (returnval){ [ 956.476414] env[63088]: value = "task-1285294" [ 956.476414] env[63088]: _type = "Task" [ 956.476414] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.485524] env[63088]: DEBUG oslo_vmware.api [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285294, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.667051] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0dd094f1-4e17-4137-b0a9-4224c597f8a7 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.840s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 956.669518] env[63088]: DEBUG oslo_concurrency.lockutils [None req-2edfcd9b-169c-4444-b2ca-db2a6c7d48b5 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.451s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 956.670062] env[63088]: DEBUG nova.objects.instance [None req-2edfcd9b-169c-4444-b2ca-db2a6c7d48b5 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lazy-loading 'resources' on Instance uuid 7cd45e52-c6be-4ec6-9710-fa99e29cf872 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 956.680263] env[63088]: DEBUG oslo_vmware.api [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Task: {'id': task-1285289, 'name': ReconfigVM_Task, 'duration_secs': 1.099353} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.681381] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] [instance: 49ee4180-6322-4003-a5b1-f2a91d190290] Reconfigured VM instance instance-00000059 to attach disk [datastore1] 49ee4180-6322-4003-a5b1-f2a91d190290/49ee4180-6322-4003-a5b1-f2a91d190290.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 956.682146] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f0b8ddc4-fcba-4a4a-aaaf-cca403fe49fa {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.691754] env[63088]: INFO nova.scheduler.client.report [None req-0dd094f1-4e17-4137-b0a9-4224c597f8a7 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Deleted allocations for instance 79c4d6e8-9999-4b0c-98d3-bbfd4d61e660 [ 956.695485] env[63088]: DEBUG oslo_vmware.api [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Waiting for the task: (returnval){ [ 956.695485] env[63088]: value = "task-1285295" [ 956.695485] env[63088]: _type = "Task" [ 956.695485] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.704865] env[63088]: INFO nova.compute.manager [-] [instance: 72c07d1e-cbb6-4875-9b7f-2649aa25eee5] Took 1.31 seconds to deallocate network for instance. [ 956.705215] env[63088]: DEBUG oslo_vmware.api [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Task: {'id': task-1285293, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.713809] env[63088]: DEBUG oslo_vmware.api [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Task: {'id': task-1285295, 'name': Rename_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.744155] env[63088]: DEBUG nova.network.neutron [-] [instance: 8205f9e1-effd-442d-990b-bf5a2a9e6cdd] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 956.802404] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a8870476-2caf-4b13-b907-85ca2b443def {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.812338] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06f8f643-2ac8-4462-8907-52b388c244ba {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.842260] env[63088]: DEBUG nova.compute.manager [req-c69bd41f-a5ca-4962-bd0b-d886d680a967 req-a2e4eeb3-ab71-4e55-9623-48ad468e5aee service nova] [instance: 8205f9e1-effd-442d-990b-bf5a2a9e6cdd] Detach interface failed, port_id=faac8e4c-2334-4c13-ae85-13e55b44fbb6, reason: Instance 8205f9e1-effd-442d-990b-bf5a2a9e6cdd could not be found. {{(pid=63088) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 956.877637] env[63088]: DEBUG oslo_concurrency.lockutils [None req-7e796b99-cf5a-4b10-9b38-5a26eafa64d7 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Lock "d6fef3ef-fec8-4929-b9b8-5e63306aa51d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.080s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 956.992024] env[63088]: DEBUG oslo_vmware.api [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285294, 'name': PowerOffVM_Task, 'duration_secs': 0.239776} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.992024] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2ad33254-8030-4454-b023-3359071f85de] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 956.992024] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2ad33254-8030-4454-b023-3359071f85de] Updating instance '2ad33254-8030-4454-b023-3359071f85de' progress to 17 {{(pid=63088) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 957.195178] env[63088]: DEBUG oslo_vmware.api [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Task: {'id': task-1285293, 'name': PowerOnVM_Task, 'duration_secs': 0.602573} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.195568] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] [instance: 4a56d74c-cb1f-4edf-99f0-4b54bf357a3a] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 957.195802] env[63088]: INFO nova.compute.manager [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] [instance: 4a56d74c-cb1f-4edf-99f0-4b54bf357a3a] Took 6.44 seconds to spawn the instance on the hypervisor. [ 957.196429] env[63088]: DEBUG nova.compute.manager [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] [instance: 4a56d74c-cb1f-4edf-99f0-4b54bf357a3a] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 957.199754] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07e94817-db29-4bb4-b21c-7084cba46ca9 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.210383] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0dd094f1-4e17-4137-b0a9-4224c597f8a7 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Lock "79c4d6e8-9999-4b0c-98d3-bbfd4d61e660" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.603s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.226362] env[63088]: DEBUG oslo_vmware.api [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Task: {'id': task-1285295, 'name': Rename_Task, 'duration_secs': 0.172985} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.227505] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] [instance: 49ee4180-6322-4003-a5b1-f2a91d190290] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 957.227863] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7c290df8-2a88-4712-82b4-9503ee7f2371 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.237178] env[63088]: DEBUG oslo_vmware.api [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Waiting for the task: (returnval){ [ 957.237178] env[63088]: value = "task-1285296" [ 957.237178] env[63088]: _type = "Task" [ 957.237178] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.247976] env[63088]: INFO nova.compute.manager [-] [instance: 8205f9e1-effd-442d-990b-bf5a2a9e6cdd] Took 1.23 seconds to deallocate network for instance. [ 957.257607] env[63088]: DEBUG oslo_vmware.api [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Task: {'id': task-1285296, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.377554] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb8b88f2-f649-4adf-94b4-2abec6bc537c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.385556] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd162c69-40c3-4f8b-bbd4-bbb611ece781 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.417693] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8de991bd-def1-4eb8-861b-89ad4d0c9694 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.425987] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eff622c7-d071-48c0-9d50-aa1426dcd661 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.440117] env[63088]: DEBUG nova.compute.provider_tree [None req-2edfcd9b-169c-4444-b2ca-db2a6c7d48b5 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 957.497973] env[63088]: DEBUG nova.virt.hardware [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:33Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 957.498371] env[63088]: DEBUG nova.virt.hardware [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 957.498539] env[63088]: DEBUG nova.virt.hardware [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 957.498943] env[63088]: DEBUG nova.virt.hardware [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 957.498943] env[63088]: DEBUG nova.virt.hardware [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 957.499095] env[63088]: DEBUG nova.virt.hardware [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 957.499294] env[63088]: DEBUG nova.virt.hardware [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 957.499463] env[63088]: DEBUG nova.virt.hardware [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 957.499639] env[63088]: DEBUG nova.virt.hardware [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 957.499868] env[63088]: DEBUG nova.virt.hardware [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 957.500070] env[63088]: DEBUG nova.virt.hardware [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 957.510033] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c6d26161-e27f-4957-9683-8111581c62ca {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.518712] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b5ff0ab0-0a14-4514-b203-4512f0d101da tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Acquiring lock "254db932-35f1-42e3-9207-cd886efd65a3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 957.518965] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b5ff0ab0-0a14-4514-b203-4512f0d101da tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Lock "254db932-35f1-42e3-9207-cd886efd65a3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 957.519180] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b5ff0ab0-0a14-4514-b203-4512f0d101da tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Acquiring lock "254db932-35f1-42e3-9207-cd886efd65a3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 957.519362] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b5ff0ab0-0a14-4514-b203-4512f0d101da tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Lock "254db932-35f1-42e3-9207-cd886efd65a3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 957.520018] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b5ff0ab0-0a14-4514-b203-4512f0d101da tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Lock "254db932-35f1-42e3-9207-cd886efd65a3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.521729] env[63088]: INFO nova.compute.manager [None req-b5ff0ab0-0a14-4514-b203-4512f0d101da tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 254db932-35f1-42e3-9207-cd886efd65a3] Terminating instance [ 957.524438] env[63088]: DEBUG nova.compute.manager [None req-b5ff0ab0-0a14-4514-b203-4512f0d101da tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 254db932-35f1-42e3-9207-cd886efd65a3] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 957.524629] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-b5ff0ab0-0a14-4514-b203-4512f0d101da tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 254db932-35f1-42e3-9207-cd886efd65a3] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 957.526505] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0254ebff-91af-4e71-aff4-f966c2e21774 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.535365] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5ff0ab0-0a14-4514-b203-4512f0d101da tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 254db932-35f1-42e3-9207-cd886efd65a3] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 957.537666] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-06094468-bf82-42d0-98b9-c3318294872d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.540386] env[63088]: DEBUG oslo_vmware.api [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Waiting for the task: (returnval){ [ 957.540386] env[63088]: value = "task-1285297" [ 957.540386] env[63088]: _type = "Task" [ 957.540386] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.548065] env[63088]: DEBUG oslo_vmware.api [None req-b5ff0ab0-0a14-4514-b203-4512f0d101da tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Waiting for the task: (returnval){ [ 957.548065] env[63088]: value = "task-1285298" [ 957.548065] env[63088]: _type = "Task" [ 957.548065] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.559613] env[63088]: DEBUG oslo_vmware.api [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285297, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.565697] env[63088]: DEBUG oslo_vmware.api [None req-b5ff0ab0-0a14-4514-b203-4512f0d101da tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285298, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.739194] env[63088]: INFO nova.compute.manager [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] [instance: 4a56d74c-cb1f-4edf-99f0-4b54bf357a3a] Took 16.13 seconds to build instance. [ 957.755757] env[63088]: DEBUG oslo_vmware.api [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Task: {'id': task-1285296, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.765671] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b3a20ade-2970-49c5-b2b9-f8e2a795f60b tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "72c07d1e-cbb6-4875-9b7f-2649aa25eee5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 2.428s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.767956] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d176fcb8-4997-4cd3-a198-d23e1a48be28 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 957.943866] env[63088]: DEBUG nova.scheduler.client.report [None req-2edfcd9b-169c-4444-b2ca-db2a6c7d48b5 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 958.054945] env[63088]: DEBUG oslo_vmware.api [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285297, 'name': ReconfigVM_Task, 'duration_secs': 0.213599} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.055801] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2ad33254-8030-4454-b023-3359071f85de] Updating instance '2ad33254-8030-4454-b023-3359071f85de' progress to 33 {{(pid=63088) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 958.063657] env[63088]: DEBUG oslo_vmware.api [None req-b5ff0ab0-0a14-4514-b203-4512f0d101da tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285298, 'name': PowerOffVM_Task, 'duration_secs': 0.256899} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.065823] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5ff0ab0-0a14-4514-b203-4512f0d101da tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 254db932-35f1-42e3-9207-cd886efd65a3] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 958.068085] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-b5ff0ab0-0a14-4514-b203-4512f0d101da tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 254db932-35f1-42e3-9207-cd886efd65a3] Unregistering the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 958.068498] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ca56f7d8-0a4d-43ba-be4e-71d086c80f41 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.137194] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-b5ff0ab0-0a14-4514-b203-4512f0d101da tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 254db932-35f1-42e3-9207-cd886efd65a3] Unregistered the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 958.137433] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-b5ff0ab0-0a14-4514-b203-4512f0d101da tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 254db932-35f1-42e3-9207-cd886efd65a3] Deleting contents of the VM from datastore datastore2 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 958.137666] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-b5ff0ab0-0a14-4514-b203-4512f0d101da tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Deleting the datastore file [datastore2] 254db932-35f1-42e3-9207-cd886efd65a3 {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 958.137954] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-edc27e13-1b31-4cb8-a79f-d49875293d1c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.149162] env[63088]: DEBUG oslo_vmware.api [None req-b5ff0ab0-0a14-4514-b203-4512f0d101da tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Waiting for the task: (returnval){ [ 958.149162] env[63088]: value = "task-1285300" [ 958.149162] env[63088]: _type = "Task" [ 958.149162] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.161466] env[63088]: DEBUG oslo_vmware.api [None req-b5ff0ab0-0a14-4514-b203-4512f0d101da tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285300, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.218576] env[63088]: DEBUG oslo_concurrency.lockutils [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Acquiring lock "818704d4-2010-4492-8e51-28ad0bd6f570" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 958.219025] env[63088]: DEBUG oslo_concurrency.lockutils [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Lock "818704d4-2010-4492-8e51-28ad0bd6f570" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.243492] env[63088]: DEBUG oslo_concurrency.lockutils [None req-58c1f530-049e-4ec5-b2a7-b3deb2afd4f4 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Lock "4a56d74c-cb1f-4edf-99f0-4b54bf357a3a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.638s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.250088] env[63088]: DEBUG oslo_vmware.api [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Task: {'id': task-1285296, 'name': PowerOnVM_Task, 'duration_secs': 0.635717} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.250812] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] [instance: 49ee4180-6322-4003-a5b1-f2a91d190290] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 958.251106] env[63088]: INFO nova.compute.manager [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] [instance: 49ee4180-6322-4003-a5b1-f2a91d190290] Took 9.86 seconds to spawn the instance on the hypervisor. [ 958.251392] env[63088]: DEBUG nova.compute.manager [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] [instance: 49ee4180-6322-4003-a5b1-f2a91d190290] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 958.252291] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c256ca48-5c2f-4195-9a5f-ab28c2812028 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.381202] env[63088]: DEBUG nova.compute.manager [req-654465a3-259d-4096-b7e5-7f698586f4e4 req-da4e052a-ece2-4929-b744-cd80fc5984a0 service nova] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Received event network-changed-81fc2dda-617e-4e4c-aa91-7e7020003995 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 958.381202] env[63088]: DEBUG nova.compute.manager [req-654465a3-259d-4096-b7e5-7f698586f4e4 req-da4e052a-ece2-4929-b744-cd80fc5984a0 service nova] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Refreshing instance network info cache due to event network-changed-81fc2dda-617e-4e4c-aa91-7e7020003995. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 958.381202] env[63088]: DEBUG oslo_concurrency.lockutils [req-654465a3-259d-4096-b7e5-7f698586f4e4 req-da4e052a-ece2-4929-b744-cd80fc5984a0 service nova] Acquiring lock "refresh_cache-f582da15-dfc3-45e1-a995-9dd0c9533869" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 958.381870] env[63088]: DEBUG oslo_concurrency.lockutils [req-654465a3-259d-4096-b7e5-7f698586f4e4 req-da4e052a-ece2-4929-b744-cd80fc5984a0 service nova] Acquired lock "refresh_cache-f582da15-dfc3-45e1-a995-9dd0c9533869" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 958.382077] env[63088]: DEBUG nova.network.neutron [req-654465a3-259d-4096-b7e5-7f698586f4e4 req-da4e052a-ece2-4929-b744-cd80fc5984a0 service nova] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Refreshing network info cache for port 81fc2dda-617e-4e4c-aa91-7e7020003995 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 958.458596] env[63088]: DEBUG oslo_concurrency.lockutils [None req-2edfcd9b-169c-4444-b2ca-db2a6c7d48b5 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.787s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.459614] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d176fcb8-4997-4cd3-a198-d23e1a48be28 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.692s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.459935] env[63088]: DEBUG nova.objects.instance [None req-d176fcb8-4997-4cd3-a198-d23e1a48be28 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Lazy-loading 'resources' on Instance uuid 8205f9e1-effd-442d-990b-bf5a2a9e6cdd {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 958.487293] env[63088]: INFO nova.scheduler.client.report [None req-2edfcd9b-169c-4444-b2ca-db2a6c7d48b5 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Deleted allocations for instance 7cd45e52-c6be-4ec6-9710-fa99e29cf872 [ 958.567953] env[63088]: DEBUG nova.virt.hardware [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 958.568806] env[63088]: DEBUG nova.virt.hardware [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 958.569152] env[63088]: DEBUG nova.virt.hardware [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 958.569541] env[63088]: DEBUG nova.virt.hardware [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 958.569860] env[63088]: DEBUG nova.virt.hardware [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 958.570228] env[63088]: DEBUG nova.virt.hardware [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 958.570630] env[63088]: DEBUG nova.virt.hardware [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 958.571159] env[63088]: DEBUG nova.virt.hardware [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 958.571400] env[63088]: DEBUG nova.virt.hardware [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 958.571646] env[63088]: DEBUG nova.virt.hardware [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 958.571901] env[63088]: DEBUG nova.virt.hardware [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 958.577574] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2ad33254-8030-4454-b023-3359071f85de] Reconfiguring VM instance instance-00000055 to detach disk 2000 {{(pid=63088) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 958.577891] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5a0c1630-1641-4eb1-9493-aa4931462fc6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.602434] env[63088]: DEBUG oslo_vmware.api [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Waiting for the task: (returnval){ [ 958.602434] env[63088]: value = "task-1285301" [ 958.602434] env[63088]: _type = "Task" [ 958.602434] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.611938] env[63088]: DEBUG oslo_vmware.api [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285301, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.645710] env[63088]: DEBUG nova.compute.manager [None req-40daf1ba-9806-483b-ace8-0ced1379d4ae tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] [instance: 4a56d74c-cb1f-4edf-99f0-4b54bf357a3a] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 958.646748] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c13f35e-5b57-4043-833f-34ecc07df730 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.662015] env[63088]: DEBUG oslo_vmware.api [None req-b5ff0ab0-0a14-4514-b203-4512f0d101da tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285300, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.174001} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.662294] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-b5ff0ab0-0a14-4514-b203-4512f0d101da tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 958.662762] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-b5ff0ab0-0a14-4514-b203-4512f0d101da tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 254db932-35f1-42e3-9207-cd886efd65a3] Deleted contents of the VM from datastore datastore2 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 958.662762] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-b5ff0ab0-0a14-4514-b203-4512f0d101da tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 254db932-35f1-42e3-9207-cd886efd65a3] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 958.662882] env[63088]: INFO nova.compute.manager [None req-b5ff0ab0-0a14-4514-b203-4512f0d101da tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 254db932-35f1-42e3-9207-cd886efd65a3] Took 1.14 seconds to destroy the instance on the hypervisor. [ 958.663137] env[63088]: DEBUG oslo.service.loopingcall [None req-b5ff0ab0-0a14-4514-b203-4512f0d101da tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 958.663567] env[63088]: DEBUG nova.compute.manager [-] [instance: 254db932-35f1-42e3-9207-cd886efd65a3] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 958.663667] env[63088]: DEBUG nova.network.neutron [-] [instance: 254db932-35f1-42e3-9207-cd886efd65a3] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 958.721889] env[63088]: DEBUG nova.compute.manager [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 818704d4-2010-4492-8e51-28ad0bd6f570] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 958.769815] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d3109219-0d50-4f6b-8903-47cb3c3c2bc0 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Acquiring lock "4a56d74c-cb1f-4edf-99f0-4b54bf357a3a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 958.770105] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d3109219-0d50-4f6b-8903-47cb3c3c2bc0 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Lock "4a56d74c-cb1f-4edf-99f0-4b54bf357a3a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.770407] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d3109219-0d50-4f6b-8903-47cb3c3c2bc0 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Acquiring lock "4a56d74c-cb1f-4edf-99f0-4b54bf357a3a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 958.770706] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d3109219-0d50-4f6b-8903-47cb3c3c2bc0 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Lock "4a56d74c-cb1f-4edf-99f0-4b54bf357a3a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.770914] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d3109219-0d50-4f6b-8903-47cb3c3c2bc0 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Lock "4a56d74c-cb1f-4edf-99f0-4b54bf357a3a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.782269] env[63088]: INFO nova.compute.manager [None req-d3109219-0d50-4f6b-8903-47cb3c3c2bc0 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] [instance: 4a56d74c-cb1f-4edf-99f0-4b54bf357a3a] Terminating instance [ 958.783870] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d3109219-0d50-4f6b-8903-47cb3c3c2bc0 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Acquiring lock "refresh_cache-4a56d74c-cb1f-4edf-99f0-4b54bf357a3a" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 958.783978] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d3109219-0d50-4f6b-8903-47cb3c3c2bc0 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Acquired lock "refresh_cache-4a56d74c-cb1f-4edf-99f0-4b54bf357a3a" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 958.784200] env[63088]: DEBUG nova.network.neutron [None req-d3109219-0d50-4f6b-8903-47cb3c3c2bc0 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] [instance: 4a56d74c-cb1f-4edf-99f0-4b54bf357a3a] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 958.787514] env[63088]: INFO nova.compute.manager [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] [instance: 49ee4180-6322-4003-a5b1-f2a91d190290] Took 20.05 seconds to build instance. [ 958.996703] env[63088]: DEBUG oslo_concurrency.lockutils [None req-2edfcd9b-169c-4444-b2ca-db2a6c7d48b5 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lock "7cd45e52-c6be-4ec6-9710-fa99e29cf872" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.881s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 959.117949] env[63088]: DEBUG oslo_vmware.api [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285301, 'name': ReconfigVM_Task, 'duration_secs': 0.306426} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.119195] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2ad33254-8030-4454-b023-3359071f85de] Reconfigured VM instance instance-00000055 to detach disk 2000 {{(pid=63088) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 959.120393] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-955dba13-b240-4d0a-9b07-25ad043765aa {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.150778] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2ad33254-8030-4454-b023-3359071f85de] Reconfiguring VM instance instance-00000055 to attach disk [datastore2] 2ad33254-8030-4454-b023-3359071f85de/2ad33254-8030-4454-b023-3359071f85de.vmdk or device None with type thin {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 959.156569] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-74e7ca2b-3472-45f4-aa9e-2589558caaf5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.174026] env[63088]: INFO nova.compute.manager [None req-40daf1ba-9806-483b-ace8-0ced1379d4ae tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] [instance: 4a56d74c-cb1f-4edf-99f0-4b54bf357a3a] instance snapshotting [ 959.174026] env[63088]: DEBUG nova.objects.instance [None req-40daf1ba-9806-483b-ace8-0ced1379d4ae tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Lazy-loading 'flavor' on Instance uuid 4a56d74c-cb1f-4edf-99f0-4b54bf357a3a {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 959.187014] env[63088]: DEBUG oslo_vmware.api [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Waiting for the task: (returnval){ [ 959.187014] env[63088]: value = "task-1285302" [ 959.187014] env[63088]: _type = "Task" [ 959.187014] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.206813] env[63088]: DEBUG oslo_vmware.api [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285302, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.215414] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71d707a1-9299-40d4-89a9-b7396a9b6df7 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.229873] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-933a0eba-9fb6-427b-8261-ffdfac0e285f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.268706] env[63088]: DEBUG oslo_concurrency.lockutils [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 959.269086] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0187809f-3dd9-466c-8a92-b2adcec3c828 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.280966] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65609610-dbf9-4053-a82a-d12a743bfd54 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.301067] env[63088]: DEBUG oslo_concurrency.lockutils [None req-2d73930f-01db-4f38-a4ea-b17399456daf tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Lock "49ee4180-6322-4003-a5b1-f2a91d190290" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.579s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 959.301968] env[63088]: DEBUG nova.compute.provider_tree [None req-d176fcb8-4997-4cd3-a198-d23e1a48be28 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 959.339703] env[63088]: DEBUG nova.network.neutron [None req-d3109219-0d50-4f6b-8903-47cb3c3c2bc0 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] [instance: 4a56d74c-cb1f-4edf-99f0-4b54bf357a3a] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 959.390731] env[63088]: DEBUG oslo_concurrency.lockutils [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquiring lock "66178bd0-7ddf-458c-98a2-3519e536dc63" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 959.390731] env[63088]: DEBUG oslo_concurrency.lockutils [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "66178bd0-7ddf-458c-98a2-3519e536dc63" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 959.449850] env[63088]: DEBUG nova.network.neutron [None req-d3109219-0d50-4f6b-8903-47cb3c3c2bc0 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] [instance: 4a56d74c-cb1f-4edf-99f0-4b54bf357a3a] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 959.512515] env[63088]: DEBUG nova.network.neutron [req-654465a3-259d-4096-b7e5-7f698586f4e4 req-da4e052a-ece2-4929-b744-cd80fc5984a0 service nova] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Updated VIF entry in instance network info cache for port 81fc2dda-617e-4e4c-aa91-7e7020003995. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 959.512926] env[63088]: DEBUG nova.network.neutron [req-654465a3-259d-4096-b7e5-7f698586f4e4 req-da4e052a-ece2-4929-b744-cd80fc5984a0 service nova] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Updating instance_info_cache with network_info: [{"id": "81fc2dda-617e-4e4c-aa91-7e7020003995", "address": "fa:16:3e:8c:09:4d", "network": {"id": "1174a284-d3ca-4f9e-aa81-13ee9a693e55", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1994276040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.150", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa3d24a1a6c0430985fd80365d986ee1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap81fc2dda-61", "ovs_interfaceid": "81fc2dda-617e-4e4c-aa91-7e7020003995", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 959.681764] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7661c8c9-b064-49dd-a13c-62d8ece70f05 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.711636] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1d0ae70-c167-4dc1-bc03-1a268830c507 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.720691] env[63088]: DEBUG oslo_vmware.api [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285302, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.774766] env[63088]: DEBUG nova.network.neutron [-] [instance: 254db932-35f1-42e3-9207-cd886efd65a3] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 959.807735] env[63088]: DEBUG nova.scheduler.client.report [None req-d176fcb8-4997-4cd3-a198-d23e1a48be28 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 959.893764] env[63088]: DEBUG nova.compute.manager [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 66178bd0-7ddf-458c-98a2-3519e536dc63] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 959.923558] env[63088]: DEBUG oslo_concurrency.lockutils [None req-093bbf73-8b98-46e9-b1d0-a1fd756e135a tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquiring lock "fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 959.923558] env[63088]: DEBUG oslo_concurrency.lockutils [None req-093bbf73-8b98-46e9-b1d0-a1fd756e135a tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lock "fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 959.923558] env[63088]: DEBUG oslo_concurrency.lockutils [None req-093bbf73-8b98-46e9-b1d0-a1fd756e135a tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquiring lock "fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 959.923558] env[63088]: DEBUG oslo_concurrency.lockutils [None req-093bbf73-8b98-46e9-b1d0-a1fd756e135a tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lock "fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 959.923558] env[63088]: DEBUG oslo_concurrency.lockutils [None req-093bbf73-8b98-46e9-b1d0-a1fd756e135a tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lock "fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 959.925032] env[63088]: INFO nova.compute.manager [None req-093bbf73-8b98-46e9-b1d0-a1fd756e135a tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3] Terminating instance [ 959.927467] env[63088]: DEBUG nova.compute.manager [None req-093bbf73-8b98-46e9-b1d0-a1fd756e135a tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 959.927848] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-093bbf73-8b98-46e9-b1d0-a1fd756e135a tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 959.929326] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5132d472-79df-46c2-8f26-43611a22c14f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.941182] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-093bbf73-8b98-46e9-b1d0-a1fd756e135a tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 959.942312] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1d5039db-66f6-42e9-84ff-0dcc7d4243da {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.953445] env[63088]: DEBUG oslo_vmware.api [None req-093bbf73-8b98-46e9-b1d0-a1fd756e135a tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Waiting for the task: (returnval){ [ 959.953445] env[63088]: value = "task-1285303" [ 959.953445] env[63088]: _type = "Task" [ 959.953445] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.954690] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d3109219-0d50-4f6b-8903-47cb3c3c2bc0 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Releasing lock "refresh_cache-4a56d74c-cb1f-4edf-99f0-4b54bf357a3a" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 959.954690] env[63088]: DEBUG nova.compute.manager [None req-d3109219-0d50-4f6b-8903-47cb3c3c2bc0 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] [instance: 4a56d74c-cb1f-4edf-99f0-4b54bf357a3a] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 959.955027] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-d3109219-0d50-4f6b-8903-47cb3c3c2bc0 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] [instance: 4a56d74c-cb1f-4edf-99f0-4b54bf357a3a] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 959.956087] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d36569cb-1663-4c34-8dbd-4249053566e2 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.972570] env[63088]: DEBUG oslo_vmware.api [None req-093bbf73-8b98-46e9-b1d0-a1fd756e135a tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285303, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.975518] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-d3109219-0d50-4f6b-8903-47cb3c3c2bc0 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] [instance: 4a56d74c-cb1f-4edf-99f0-4b54bf357a3a] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 959.975880] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-76eda1dd-850e-47d4-988b-3c9da8a8135d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.983897] env[63088]: DEBUG oslo_vmware.api [None req-d3109219-0d50-4f6b-8903-47cb3c3c2bc0 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Waiting for the task: (returnval){ [ 959.983897] env[63088]: value = "task-1285304" [ 959.983897] env[63088]: _type = "Task" [ 959.983897] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.994407] env[63088]: DEBUG oslo_vmware.api [None req-d3109219-0d50-4f6b-8903-47cb3c3c2bc0 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Task: {'id': task-1285304, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.016463] env[63088]: DEBUG oslo_concurrency.lockutils [req-654465a3-259d-4096-b7e5-7f698586f4e4 req-da4e052a-ece2-4929-b744-cd80fc5984a0 service nova] Releasing lock "refresh_cache-f582da15-dfc3-45e1-a995-9dd0c9533869" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 960.017032] env[63088]: DEBUG nova.compute.manager [req-654465a3-259d-4096-b7e5-7f698586f4e4 req-da4e052a-ece2-4929-b744-cd80fc5984a0 service nova] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Received event network-changed-b7a01048-9168-46ff-a5b8-83cee3dd3c97 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 960.017032] env[63088]: DEBUG nova.compute.manager [req-654465a3-259d-4096-b7e5-7f698586f4e4 req-da4e052a-ece2-4929-b744-cd80fc5984a0 service nova] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Refreshing instance network info cache due to event network-changed-b7a01048-9168-46ff-a5b8-83cee3dd3c97. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 960.017159] env[63088]: DEBUG oslo_concurrency.lockutils [req-654465a3-259d-4096-b7e5-7f698586f4e4 req-da4e052a-ece2-4929-b744-cd80fc5984a0 service nova] Acquiring lock "refresh_cache-d6fef3ef-fec8-4929-b9b8-5e63306aa51d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 960.017329] env[63088]: DEBUG oslo_concurrency.lockutils [req-654465a3-259d-4096-b7e5-7f698586f4e4 req-da4e052a-ece2-4929-b744-cd80fc5984a0 service nova] Acquired lock "refresh_cache-d6fef3ef-fec8-4929-b9b8-5e63306aa51d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 960.017532] env[63088]: DEBUG nova.network.neutron [req-654465a3-259d-4096-b7e5-7f698586f4e4 req-da4e052a-ece2-4929-b744-cd80fc5984a0 service nova] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Refreshing network info cache for port b7a01048-9168-46ff-a5b8-83cee3dd3c97 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 960.198968] env[63088]: DEBUG oslo_vmware.api [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285302, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.231845] env[63088]: DEBUG nova.compute.manager [None req-40daf1ba-9806-483b-ace8-0ced1379d4ae tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] [instance: 4a56d74c-cb1f-4edf-99f0-4b54bf357a3a] Instance disappeared during snapshot {{(pid=63088) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4494}} [ 960.276912] env[63088]: INFO nova.compute.manager [-] [instance: 254db932-35f1-42e3-9207-cd886efd65a3] Took 1.61 seconds to deallocate network for instance. [ 960.312837] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d176fcb8-4997-4cd3-a198-d23e1a48be28 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.854s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.316259] env[63088]: DEBUG oslo_concurrency.lockutils [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.048s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.317736] env[63088]: INFO nova.compute.claims [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 818704d4-2010-4492-8e51-28ad0bd6f570] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 960.336701] env[63088]: INFO nova.scheduler.client.report [None req-d176fcb8-4997-4cd3-a198-d23e1a48be28 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Deleted allocations for instance 8205f9e1-effd-442d-990b-bf5a2a9e6cdd [ 960.366390] env[63088]: DEBUG nova.compute.manager [None req-40daf1ba-9806-483b-ace8-0ced1379d4ae tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] [instance: 4a56d74c-cb1f-4edf-99f0-4b54bf357a3a] Found 0 images (rotation: 2) {{(pid=63088) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 960.414296] env[63088]: DEBUG nova.compute.manager [req-d4fa64a4-4255-4d4c-95fd-f1f9c9d95bbf req-1bf75170-17ba-4148-965d-f6f64ff781b0 service nova] [instance: 254db932-35f1-42e3-9207-cd886efd65a3] Received event network-vif-deleted-d2722fc9-4b54-495f-81bb-d5b411285079 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 960.414575] env[63088]: DEBUG nova.compute.manager [req-d4fa64a4-4255-4d4c-95fd-f1f9c9d95bbf req-1bf75170-17ba-4148-965d-f6f64ff781b0 service nova] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Received event network-changed-b7a01048-9168-46ff-a5b8-83cee3dd3c97 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 960.414763] env[63088]: DEBUG nova.compute.manager [req-d4fa64a4-4255-4d4c-95fd-f1f9c9d95bbf req-1bf75170-17ba-4148-965d-f6f64ff781b0 service nova] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Refreshing instance network info cache due to event network-changed-b7a01048-9168-46ff-a5b8-83cee3dd3c97. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 960.414904] env[63088]: DEBUG oslo_concurrency.lockutils [req-d4fa64a4-4255-4d4c-95fd-f1f9c9d95bbf req-1bf75170-17ba-4148-965d-f6f64ff781b0 service nova] Acquiring lock "refresh_cache-d6fef3ef-fec8-4929-b9b8-5e63306aa51d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 960.415914] env[63088]: DEBUG oslo_concurrency.lockutils [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.466276] env[63088]: DEBUG oslo_vmware.api [None req-093bbf73-8b98-46e9-b1d0-a1fd756e135a tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285303, 'name': PowerOffVM_Task, 'duration_secs': 0.20696} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.466547] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-093bbf73-8b98-46e9-b1d0-a1fd756e135a tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 960.466726] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-093bbf73-8b98-46e9-b1d0-a1fd756e135a tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3] Unregistering the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 960.467033] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-31738673-b93e-4c04-bb63-665fd7f63d8c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.494147] env[63088]: DEBUG oslo_vmware.api [None req-d3109219-0d50-4f6b-8903-47cb3c3c2bc0 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Task: {'id': task-1285304, 'name': PowerOffVM_Task, 'duration_secs': 0.153576} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.494448] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-d3109219-0d50-4f6b-8903-47cb3c3c2bc0 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] [instance: 4a56d74c-cb1f-4edf-99f0-4b54bf357a3a] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 960.494634] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-d3109219-0d50-4f6b-8903-47cb3c3c2bc0 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] [instance: 4a56d74c-cb1f-4edf-99f0-4b54bf357a3a] Unregistering the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 960.494886] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3f0456b4-07d3-4968-966c-684a63a02e0c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.522144] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-d3109219-0d50-4f6b-8903-47cb3c3c2bc0 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] [instance: 4a56d74c-cb1f-4edf-99f0-4b54bf357a3a] Unregistered the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 960.522497] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-d3109219-0d50-4f6b-8903-47cb3c3c2bc0 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] [instance: 4a56d74c-cb1f-4edf-99f0-4b54bf357a3a] Deleting contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 960.522739] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-d3109219-0d50-4f6b-8903-47cb3c3c2bc0 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Deleting the datastore file [datastore1] 4a56d74c-cb1f-4edf-99f0-4b54bf357a3a {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 960.525266] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a6622018-576f-4fec-b257-a0dbafc3218c tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Acquiring lock "49ee4180-6322-4003-a5b1-f2a91d190290" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.525463] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a6622018-576f-4fec-b257-a0dbafc3218c tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Lock "49ee4180-6322-4003-a5b1-f2a91d190290" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.525722] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a6622018-576f-4fec-b257-a0dbafc3218c tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Acquiring lock "49ee4180-6322-4003-a5b1-f2a91d190290-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.526012] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a6622018-576f-4fec-b257-a0dbafc3218c tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Lock "49ee4180-6322-4003-a5b1-f2a91d190290-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.526221] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a6622018-576f-4fec-b257-a0dbafc3218c tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Lock "49ee4180-6322-4003-a5b1-f2a91d190290-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.528061] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d34bd5e3-de25-4507-b855-a189e0345dcf {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.531317] env[63088]: INFO nova.compute.manager [None req-a6622018-576f-4fec-b257-a0dbafc3218c tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] [instance: 49ee4180-6322-4003-a5b1-f2a91d190290] Terminating instance [ 960.533983] env[63088]: DEBUG nova.compute.manager [None req-a6622018-576f-4fec-b257-a0dbafc3218c tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] [instance: 49ee4180-6322-4003-a5b1-f2a91d190290] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 960.534207] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-a6622018-576f-4fec-b257-a0dbafc3218c tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] [instance: 49ee4180-6322-4003-a5b1-f2a91d190290] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 960.535133] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf53bf56-2bb3-401a-afb7-256ed4aaded8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.539214] env[63088]: DEBUG oslo_vmware.api [None req-d3109219-0d50-4f6b-8903-47cb3c3c2bc0 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Waiting for the task: (returnval){ [ 960.539214] env[63088]: value = "task-1285307" [ 960.539214] env[63088]: _type = "Task" [ 960.539214] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.546370] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6622018-576f-4fec-b257-a0dbafc3218c tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] [instance: 49ee4180-6322-4003-a5b1-f2a91d190290] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 960.547102] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-093bbf73-8b98-46e9-b1d0-a1fd756e135a tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3] Unregistered the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 960.547319] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-093bbf73-8b98-46e9-b1d0-a1fd756e135a tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3] Deleting contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 960.547502] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-093bbf73-8b98-46e9-b1d0-a1fd756e135a tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Deleting the datastore file [datastore1] fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3 {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 960.548128] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-295b2027-7661-45ac-b535-38edc243b932 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.549694] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e3365c3f-0544-4387-b168-40097883b012 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.554751] env[63088]: DEBUG oslo_vmware.api [None req-d3109219-0d50-4f6b-8903-47cb3c3c2bc0 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Task: {'id': task-1285307, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.561575] env[63088]: DEBUG oslo_vmware.api [None req-093bbf73-8b98-46e9-b1d0-a1fd756e135a tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Waiting for the task: (returnval){ [ 960.561575] env[63088]: value = "task-1285309" [ 960.561575] env[63088]: _type = "Task" [ 960.561575] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.562640] env[63088]: DEBUG oslo_vmware.api [None req-a6622018-576f-4fec-b257-a0dbafc3218c tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Waiting for the task: (returnval){ [ 960.562640] env[63088]: value = "task-1285308" [ 960.562640] env[63088]: _type = "Task" [ 960.562640] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.577540] env[63088]: DEBUG oslo_vmware.api [None req-093bbf73-8b98-46e9-b1d0-a1fd756e135a tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285309, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.580748] env[63088]: DEBUG oslo_vmware.api [None req-a6622018-576f-4fec-b257-a0dbafc3218c tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Task: {'id': task-1285308, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.701838] env[63088]: DEBUG oslo_vmware.api [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285302, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.783441] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b5ff0ab0-0a14-4514-b203-4512f0d101da tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.834897] env[63088]: DEBUG nova.network.neutron [req-654465a3-259d-4096-b7e5-7f698586f4e4 req-da4e052a-ece2-4929-b744-cd80fc5984a0 service nova] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Updated VIF entry in instance network info cache for port b7a01048-9168-46ff-a5b8-83cee3dd3c97. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 960.835355] env[63088]: DEBUG nova.network.neutron [req-654465a3-259d-4096-b7e5-7f698586f4e4 req-da4e052a-ece2-4929-b744-cd80fc5984a0 service nova] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Updating instance_info_cache with network_info: [{"id": "b7a01048-9168-46ff-a5b8-83cee3dd3c97", "address": "fa:16:3e:e3:57:91", "network": {"id": "1174a284-d3ca-4f9e-aa81-13ee9a693e55", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1994276040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa3d24a1a6c0430985fd80365d986ee1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb7a01048-91", "ovs_interfaceid": "b7a01048-9168-46ff-a5b8-83cee3dd3c97", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 960.847422] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d176fcb8-4997-4cd3-a198-d23e1a48be28 tempest-ServerAddressesTestJSON-921133807 tempest-ServerAddressesTestJSON-921133807-project-member] Lock "8205f9e1-effd-442d-990b-bf5a2a9e6cdd" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.924374] env[63088]: DEBUG oslo_vmware.rw_handles [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/521c4025-547d-e28f-7ab0-4f22d06e84bc/disk-0.vmdk. {{(pid=63088) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 960.925686] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7164a802-db94-4873-8cbe-16411e54d258 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.943326] env[63088]: DEBUG oslo_vmware.rw_handles [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/521c4025-547d-e28f-7ab0-4f22d06e84bc/disk-0.vmdk is in state: ready. {{(pid=63088) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 960.943588] env[63088]: ERROR oslo_vmware.rw_handles [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/521c4025-547d-e28f-7ab0-4f22d06e84bc/disk-0.vmdk due to incomplete transfer. [ 960.943844] env[63088]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-53b7ab9e-4f8c-4305-9d76-a6e86f8acd65 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.952826] env[63088]: DEBUG oslo_vmware.rw_handles [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/521c4025-547d-e28f-7ab0-4f22d06e84bc/disk-0.vmdk. {{(pid=63088) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 960.953096] env[63088]: DEBUG nova.virt.vmwareapi.images [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Uploaded image e3652b5e-39c5-4fe1-8b63-944e4527acd9 to the Glance image server {{(pid=63088) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 960.955802] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Destroying the VM {{(pid=63088) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 960.956153] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-6d1edef7-b394-47ad-96c7-c26a743ff43a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.964804] env[63088]: DEBUG oslo_vmware.api [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Waiting for the task: (returnval){ [ 960.964804] env[63088]: value = "task-1285310" [ 960.964804] env[63088]: _type = "Task" [ 960.964804] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.975724] env[63088]: DEBUG oslo_vmware.api [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285310, 'name': Destroy_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.051588] env[63088]: DEBUG oslo_vmware.api [None req-d3109219-0d50-4f6b-8903-47cb3c3c2bc0 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Task: {'id': task-1285307, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.108421} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.051588] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-d3109219-0d50-4f6b-8903-47cb3c3c2bc0 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 961.051989] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-d3109219-0d50-4f6b-8903-47cb3c3c2bc0 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] [instance: 4a56d74c-cb1f-4edf-99f0-4b54bf357a3a] Deleted contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 961.051989] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-d3109219-0d50-4f6b-8903-47cb3c3c2bc0 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] [instance: 4a56d74c-cb1f-4edf-99f0-4b54bf357a3a] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 961.052112] env[63088]: INFO nova.compute.manager [None req-d3109219-0d50-4f6b-8903-47cb3c3c2bc0 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] [instance: 4a56d74c-cb1f-4edf-99f0-4b54bf357a3a] Took 1.10 seconds to destroy the instance on the hypervisor. [ 961.052451] env[63088]: DEBUG oslo.service.loopingcall [None req-d3109219-0d50-4f6b-8903-47cb3c3c2bc0 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 961.053479] env[63088]: DEBUG nova.compute.manager [-] [instance: 4a56d74c-cb1f-4edf-99f0-4b54bf357a3a] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 961.053606] env[63088]: DEBUG nova.network.neutron [-] [instance: 4a56d74c-cb1f-4edf-99f0-4b54bf357a3a] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 961.073322] env[63088]: DEBUG oslo_vmware.api [None req-093bbf73-8b98-46e9-b1d0-a1fd756e135a tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285309, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.180045} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.076296] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-093bbf73-8b98-46e9-b1d0-a1fd756e135a tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 961.076485] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-093bbf73-8b98-46e9-b1d0-a1fd756e135a tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3] Deleted contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 961.076714] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-093bbf73-8b98-46e9-b1d0-a1fd756e135a tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 961.076921] env[63088]: INFO nova.compute.manager [None req-093bbf73-8b98-46e9-b1d0-a1fd756e135a tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3] Took 1.15 seconds to destroy the instance on the hypervisor. [ 961.077194] env[63088]: DEBUG oslo.service.loopingcall [None req-093bbf73-8b98-46e9-b1d0-a1fd756e135a tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 961.077401] env[63088]: DEBUG nova.compute.manager [-] [instance: fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 961.077513] env[63088]: DEBUG nova.network.neutron [-] [instance: fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 961.084650] env[63088]: DEBUG oslo_vmware.api [None req-a6622018-576f-4fec-b257-a0dbafc3218c tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Task: {'id': task-1285308, 'name': PowerOffVM_Task, 'duration_secs': 0.248895} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.085326] env[63088]: DEBUG nova.network.neutron [-] [instance: 4a56d74c-cb1f-4edf-99f0-4b54bf357a3a] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 961.086525] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6622018-576f-4fec-b257-a0dbafc3218c tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] [instance: 49ee4180-6322-4003-a5b1-f2a91d190290] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 961.087301] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-a6622018-576f-4fec-b257-a0dbafc3218c tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] [instance: 49ee4180-6322-4003-a5b1-f2a91d190290] Unregistering the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 961.087301] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b4bf4443-cee1-4da5-8fe1-e3d9da26616c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.209079] env[63088]: DEBUG oslo_vmware.api [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285302, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.338571] env[63088]: DEBUG oslo_concurrency.lockutils [req-654465a3-259d-4096-b7e5-7f698586f4e4 req-da4e052a-ece2-4929-b744-cd80fc5984a0 service nova] Releasing lock "refresh_cache-d6fef3ef-fec8-4929-b9b8-5e63306aa51d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 961.341306] env[63088]: DEBUG oslo_concurrency.lockutils [req-d4fa64a4-4255-4d4c-95fd-f1f9c9d95bbf req-1bf75170-17ba-4148-965d-f6f64ff781b0 service nova] Acquired lock "refresh_cache-d6fef3ef-fec8-4929-b9b8-5e63306aa51d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 961.341306] env[63088]: DEBUG nova.network.neutron [req-d4fa64a4-4255-4d4c-95fd-f1f9c9d95bbf req-1bf75170-17ba-4148-965d-f6f64ff781b0 service nova] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Refreshing network info cache for port b7a01048-9168-46ff-a5b8-83cee3dd3c97 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 961.482239] env[63088]: DEBUG oslo_vmware.api [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285310, 'name': Destroy_Task} progress is 33%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.484370] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-510ce9da-abe1-4089-adf1-81f7d63bc46d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.494816] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0512e021-e790-4f4e-bda6-66e77119cf9a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.542775] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c33640a6-f377-4325-ae11-d49cd4ec6582 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.554586] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9f1c052-dbc4-48f0-b1a7-a8cd2d17b5a7 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.575639] env[63088]: DEBUG nova.compute.provider_tree [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 961.588226] env[63088]: DEBUG nova.network.neutron [-] [instance: 4a56d74c-cb1f-4edf-99f0-4b54bf357a3a] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 961.701244] env[63088]: DEBUG oslo_vmware.api [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285302, 'name': ReconfigVM_Task, 'duration_secs': 2.126592} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.701594] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2ad33254-8030-4454-b023-3359071f85de] Reconfigured VM instance instance-00000055 to attach disk [datastore2] 2ad33254-8030-4454-b023-3359071f85de/2ad33254-8030-4454-b023-3359071f85de.vmdk or device None with type thin {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 961.701914] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2ad33254-8030-4454-b023-3359071f85de] Updating instance '2ad33254-8030-4454-b023-3359071f85de' progress to 50 {{(pid=63088) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 961.733726] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-a6622018-576f-4fec-b257-a0dbafc3218c tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] [instance: 49ee4180-6322-4003-a5b1-f2a91d190290] Unregistered the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 961.733966] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-a6622018-576f-4fec-b257-a0dbafc3218c tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] [instance: 49ee4180-6322-4003-a5b1-f2a91d190290] Deleting contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 961.734249] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-a6622018-576f-4fec-b257-a0dbafc3218c tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Deleting the datastore file [datastore1] 49ee4180-6322-4003-a5b1-f2a91d190290 {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 961.734503] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e67fdd3a-6863-4ca1-b64a-dcf2b5169e29 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.741992] env[63088]: DEBUG oslo_vmware.api [None req-a6622018-576f-4fec-b257-a0dbafc3218c tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Waiting for the task: (returnval){ [ 961.741992] env[63088]: value = "task-1285312" [ 961.741992] env[63088]: _type = "Task" [ 961.741992] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.751362] env[63088]: DEBUG oslo_vmware.api [None req-a6622018-576f-4fec-b257-a0dbafc3218c tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Task: {'id': task-1285312, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.975624] env[63088]: DEBUG oslo_vmware.api [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285310, 'name': Destroy_Task, 'duration_secs': 0.882288} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.976092] env[63088]: INFO nova.virt.vmwareapi.vm_util [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Destroyed the VM [ 961.976405] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Deleting Snapshot of the VM instance {{(pid=63088) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 961.976708] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-32075cc8-d11e-4186-a8d2-9b7633954b35 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.986325] env[63088]: DEBUG oslo_vmware.api [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Waiting for the task: (returnval){ [ 961.986325] env[63088]: value = "task-1285313" [ 961.986325] env[63088]: _type = "Task" [ 961.986325] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.995414] env[63088]: DEBUG oslo_vmware.api [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285313, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.078688] env[63088]: DEBUG nova.scheduler.client.report [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 962.091290] env[63088]: INFO nova.compute.manager [-] [instance: 4a56d74c-cb1f-4edf-99f0-4b54bf357a3a] Took 1.04 seconds to deallocate network for instance. [ 962.184537] env[63088]: DEBUG nova.network.neutron [-] [instance: fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 962.210387] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19925f2b-4d0c-40a8-9126-8b67507d981d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.235326] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35319164-7094-4e83-9770-6707c474138a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.254596] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2ad33254-8030-4454-b023-3359071f85de] Updating instance '2ad33254-8030-4454-b023-3359071f85de' progress to 67 {{(pid=63088) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 962.258731] env[63088]: DEBUG nova.network.neutron [req-d4fa64a4-4255-4d4c-95fd-f1f9c9d95bbf req-1bf75170-17ba-4148-965d-f6f64ff781b0 service nova] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Updated VIF entry in instance network info cache for port b7a01048-9168-46ff-a5b8-83cee3dd3c97. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 962.259085] env[63088]: DEBUG nova.network.neutron [req-d4fa64a4-4255-4d4c-95fd-f1f9c9d95bbf req-1bf75170-17ba-4148-965d-f6f64ff781b0 service nova] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Updating instance_info_cache with network_info: [{"id": "b7a01048-9168-46ff-a5b8-83cee3dd3c97", "address": "fa:16:3e:e3:57:91", "network": {"id": "1174a284-d3ca-4f9e-aa81-13ee9a693e55", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1994276040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa3d24a1a6c0430985fd80365d986ee1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb7a01048-91", "ovs_interfaceid": "b7a01048-9168-46ff-a5b8-83cee3dd3c97", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 962.270331] env[63088]: DEBUG oslo_vmware.api [None req-a6622018-576f-4fec-b257-a0dbafc3218c tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Task: {'id': task-1285312, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.372256} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.271097] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-a6622018-576f-4fec-b257-a0dbafc3218c tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 962.271322] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-a6622018-576f-4fec-b257-a0dbafc3218c tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] [instance: 49ee4180-6322-4003-a5b1-f2a91d190290] Deleted contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 962.271513] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-a6622018-576f-4fec-b257-a0dbafc3218c tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] [instance: 49ee4180-6322-4003-a5b1-f2a91d190290] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 962.271683] env[63088]: INFO nova.compute.manager [None req-a6622018-576f-4fec-b257-a0dbafc3218c tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] [instance: 49ee4180-6322-4003-a5b1-f2a91d190290] Took 1.74 seconds to destroy the instance on the hypervisor. [ 962.271922] env[63088]: DEBUG oslo.service.loopingcall [None req-a6622018-576f-4fec-b257-a0dbafc3218c tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 962.272147] env[63088]: DEBUG nova.compute.manager [-] [instance: 49ee4180-6322-4003-a5b1-f2a91d190290] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 962.272278] env[63088]: DEBUG nova.network.neutron [-] [instance: 49ee4180-6322-4003-a5b1-f2a91d190290] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 962.443827] env[63088]: DEBUG nova.compute.manager [req-701c6726-7de5-4e96-8c48-6d22944887db req-cc7ffe7a-0675-497c-8a0b-f71c8c76b6d3 service nova] [instance: fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3] Received event network-vif-deleted-7e7a24aa-ad61-4e47-a533-6be5f25f7a37 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 962.497305] env[63088]: DEBUG oslo_vmware.api [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285313, 'name': RemoveSnapshot_Task, 'duration_secs': 0.37572} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.497605] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Deleted Snapshot of the VM instance {{(pid=63088) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 962.497887] env[63088]: DEBUG nova.compute.manager [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 962.498738] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-627525b3-6862-4c9d-be08-1910a0494ba0 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.583502] env[63088]: DEBUG oslo_concurrency.lockutils [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.267s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.584022] env[63088]: DEBUG nova.compute.manager [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 818704d4-2010-4492-8e51-28ad0bd6f570] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 962.586797] env[63088]: DEBUG oslo_concurrency.lockutils [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.171s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 962.589298] env[63088]: INFO nova.compute.claims [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 66178bd0-7ddf-458c-98a2-3519e536dc63] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 962.599043] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d3109219-0d50-4f6b-8903-47cb3c3c2bc0 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 962.687169] env[63088]: INFO nova.compute.manager [-] [instance: fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3] Took 1.61 seconds to deallocate network for instance. [ 962.764721] env[63088]: DEBUG oslo_concurrency.lockutils [req-d4fa64a4-4255-4d4c-95fd-f1f9c9d95bbf req-1bf75170-17ba-4148-965d-f6f64ff781b0 service nova] Releasing lock "refresh_cache-d6fef3ef-fec8-4929-b9b8-5e63306aa51d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 962.765012] env[63088]: DEBUG nova.compute.manager [req-d4fa64a4-4255-4d4c-95fd-f1f9c9d95bbf req-1bf75170-17ba-4148-965d-f6f64ff781b0 service nova] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Received event network-changed-81fc2dda-617e-4e4c-aa91-7e7020003995 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 962.765205] env[63088]: DEBUG nova.compute.manager [req-d4fa64a4-4255-4d4c-95fd-f1f9c9d95bbf req-1bf75170-17ba-4148-965d-f6f64ff781b0 service nova] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Refreshing instance network info cache due to event network-changed-81fc2dda-617e-4e4c-aa91-7e7020003995. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 962.765417] env[63088]: DEBUG oslo_concurrency.lockutils [req-d4fa64a4-4255-4d4c-95fd-f1f9c9d95bbf req-1bf75170-17ba-4148-965d-f6f64ff781b0 service nova] Acquiring lock "refresh_cache-f582da15-dfc3-45e1-a995-9dd0c9533869" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 962.765562] env[63088]: DEBUG oslo_concurrency.lockutils [req-d4fa64a4-4255-4d4c-95fd-f1f9c9d95bbf req-1bf75170-17ba-4148-965d-f6f64ff781b0 service nova] Acquired lock "refresh_cache-f582da15-dfc3-45e1-a995-9dd0c9533869" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 962.765724] env[63088]: DEBUG nova.network.neutron [req-d4fa64a4-4255-4d4c-95fd-f1f9c9d95bbf req-1bf75170-17ba-4148-965d-f6f64ff781b0 service nova] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Refreshing network info cache for port 81fc2dda-617e-4e4c-aa91-7e7020003995 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 962.824498] env[63088]: DEBUG nova.network.neutron [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2ad33254-8030-4454-b023-3359071f85de] Port 1d2bee7e-bbc7-4b17-976b-47e7bbb268af binding to destination host cpu-1 is already ACTIVE {{(pid=63088) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 963.013540] env[63088]: INFO nova.compute.manager [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Shelve offloading [ 963.015195] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 963.015454] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0159717e-c63b-437d-8c3d-2b2c23786e60 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.025526] env[63088]: DEBUG oslo_vmware.api [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Waiting for the task: (returnval){ [ 963.025526] env[63088]: value = "task-1285314" [ 963.025526] env[63088]: _type = "Task" [ 963.025526] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.034623] env[63088]: DEBUG oslo_vmware.api [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285314, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.034950] env[63088]: DEBUG nova.network.neutron [-] [instance: 49ee4180-6322-4003-a5b1-f2a91d190290] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 963.092916] env[63088]: DEBUG nova.compute.utils [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 963.096426] env[63088]: DEBUG nova.compute.manager [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 818704d4-2010-4492-8e51-28ad0bd6f570] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 963.096426] env[63088]: DEBUG nova.network.neutron [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 818704d4-2010-4492-8e51-28ad0bd6f570] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 963.139523] env[63088]: DEBUG nova.policy [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a6f91fc8922c42f985f6f6338dab0708', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7bff1382b9694df08133c88a5fe783a5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 963.194605] env[63088]: DEBUG oslo_concurrency.lockutils [None req-093bbf73-8b98-46e9-b1d0-a1fd756e135a tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 963.538599] env[63088]: INFO nova.compute.manager [-] [instance: 49ee4180-6322-4003-a5b1-f2a91d190290] Took 1.27 seconds to deallocate network for instance. [ 963.541029] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] VM already powered off {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 963.541295] env[63088]: DEBUG nova.compute.manager [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 963.545102] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0195555-309f-438c-b62b-d497db516c12 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.554040] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Acquiring lock "refresh_cache-71e37d8e-a454-46c4-a3cc-3d5671a32beb" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 963.554218] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Acquired lock "refresh_cache-71e37d8e-a454-46c4-a3cc-3d5671a32beb" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 963.554414] env[63088]: DEBUG nova.network.neutron [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 963.596817] env[63088]: DEBUG nova.compute.manager [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 818704d4-2010-4492-8e51-28ad0bd6f570] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 963.638705] env[63088]: DEBUG nova.network.neutron [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 818704d4-2010-4492-8e51-28ad0bd6f570] Successfully created port: 2a414d49-0028-4cbf-bc84-04946140a76c {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 963.698328] env[63088]: DEBUG nova.network.neutron [req-d4fa64a4-4255-4d4c-95fd-f1f9c9d95bbf req-1bf75170-17ba-4148-965d-f6f64ff781b0 service nova] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Updated VIF entry in instance network info cache for port 81fc2dda-617e-4e4c-aa91-7e7020003995. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 963.699022] env[63088]: DEBUG nova.network.neutron [req-d4fa64a4-4255-4d4c-95fd-f1f9c9d95bbf req-1bf75170-17ba-4148-965d-f6f64ff781b0 service nova] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Updating instance_info_cache with network_info: [{"id": "81fc2dda-617e-4e4c-aa91-7e7020003995", "address": "fa:16:3e:8c:09:4d", "network": {"id": "1174a284-d3ca-4f9e-aa81-13ee9a693e55", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1994276040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.150", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa3d24a1a6c0430985fd80365d986ee1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap81fc2dda-61", "ovs_interfaceid": "81fc2dda-617e-4e4c-aa91-7e7020003995", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 963.780800] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80b8e824-e61a-43df-923a-9b88c733a759 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.789607] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-169c663f-d320-46fc-8441-73680d49becd {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.821847] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54762b54-b6d2-43cd-a32d-d50cb78311c8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.828291] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1481a1d-f065-4583-b9f4-5c64e84dedf8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.850335] env[63088]: DEBUG oslo_concurrency.lockutils [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquiring lock "2ad33254-8030-4454-b023-3359071f85de-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 963.850582] env[63088]: DEBUG oslo_concurrency.lockutils [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Lock "2ad33254-8030-4454-b023-3359071f85de-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 963.850759] env[63088]: DEBUG oslo_concurrency.lockutils [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Lock "2ad33254-8030-4454-b023-3359071f85de-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.852055] env[63088]: DEBUG nova.compute.provider_tree [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 964.050959] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a6622018-576f-4fec-b257-a0dbafc3218c tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.201542] env[63088]: DEBUG oslo_concurrency.lockutils [req-d4fa64a4-4255-4d4c-95fd-f1f9c9d95bbf req-1bf75170-17ba-4148-965d-f6f64ff781b0 service nova] Releasing lock "refresh_cache-f582da15-dfc3-45e1-a995-9dd0c9533869" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 964.316066] env[63088]: DEBUG nova.network.neutron [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Updating instance_info_cache with network_info: [{"id": "79ca8279-a3e2-4792-a4b5-57f899be5c41", "address": "fa:16:3e:53:f8:3c", "network": {"id": "cc2c0304-8906-443a-a76e-e7a32158346f", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1867758042-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.139", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b18ddddf2314d83addf550b8cb91977", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0f096917-a0cf-4add-a9d2-23ca1c723b3b", "external-id": "nsx-vlan-transportzone-894", "segmentation_id": 894, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap79ca8279-a3", "ovs_interfaceid": "79ca8279-a3e2-4792-a4b5-57f899be5c41", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 964.362023] env[63088]: DEBUG nova.scheduler.client.report [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 964.470925] env[63088]: DEBUG nova.compute.manager [req-b3fcae0c-c3e2-4cc0-a131-9d4c4665baf8 req-f61d71b1-24ae-4830-b0fb-4617e406af82 service nova] [instance: 49ee4180-6322-4003-a5b1-f2a91d190290] Received event network-vif-deleted-15b6a284-5552-4594-97bf-23b7ec018f5d {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 964.610782] env[63088]: DEBUG nova.compute.manager [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 818704d4-2010-4492-8e51-28ad0bd6f570] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 964.636200] env[63088]: DEBUG nova.virt.hardware [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 964.636490] env[63088]: DEBUG nova.virt.hardware [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 964.636677] env[63088]: DEBUG nova.virt.hardware [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 964.636864] env[63088]: DEBUG nova.virt.hardware [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 964.637027] env[63088]: DEBUG nova.virt.hardware [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 964.637188] env[63088]: DEBUG nova.virt.hardware [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 964.637398] env[63088]: DEBUG nova.virt.hardware [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 964.637560] env[63088]: DEBUG nova.virt.hardware [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 964.637729] env[63088]: DEBUG nova.virt.hardware [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 964.637891] env[63088]: DEBUG nova.virt.hardware [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 964.638075] env[63088]: DEBUG nova.virt.hardware [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 964.639469] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4246bce-5905-446b-8fa6-cc07f0bfbee5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.648436] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d844f5e-5e34-410d-81e8-1dcd2014aa27 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.818544] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Releasing lock "refresh_cache-71e37d8e-a454-46c4-a3cc-3d5671a32beb" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 964.868657] env[63088]: DEBUG oslo_concurrency.lockutils [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.282s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.869068] env[63088]: DEBUG nova.compute.manager [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 66178bd0-7ddf-458c-98a2-3519e536dc63] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 964.871716] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b5ff0ab0-0a14-4514-b203-4512f0d101da tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.088s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.871875] env[63088]: DEBUG nova.objects.instance [None req-b5ff0ab0-0a14-4514-b203-4512f0d101da tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Lazy-loading 'resources' on Instance uuid 254db932-35f1-42e3-9207-cd886efd65a3 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 964.913980] env[63088]: DEBUG oslo_concurrency.lockutils [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquiring lock "refresh_cache-2ad33254-8030-4454-b023-3359071f85de" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 964.914370] env[63088]: DEBUG oslo_concurrency.lockutils [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquired lock "refresh_cache-2ad33254-8030-4454-b023-3359071f85de" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 964.914573] env[63088]: DEBUG nova.network.neutron [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2ad33254-8030-4454-b023-3359071f85de] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 965.333132] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 965.334086] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd9fd023-73d0-40ff-98b5-48011b5f26a8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.343100] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Unregistering the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 965.343371] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-41199a55-ca9a-4f6e-a3c0-48f071f309b4 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.374900] env[63088]: DEBUG nova.compute.utils [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 965.379817] env[63088]: DEBUG nova.compute.manager [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 66178bd0-7ddf-458c-98a2-3519e536dc63] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 965.380060] env[63088]: DEBUG nova.network.neutron [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 66178bd0-7ddf-458c-98a2-3519e536dc63] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 965.407644] env[63088]: DEBUG nova.network.neutron [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 818704d4-2010-4492-8e51-28ad0bd6f570] Successfully updated port: 2a414d49-0028-4cbf-bc84-04946140a76c {{(pid=63088) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 965.435051] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Unregistered the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 965.435460] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Deleting contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 965.435460] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Deleting the datastore file [datastore1] 71e37d8e-a454-46c4-a3cc-3d5671a32beb {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 965.436714] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-80edc2c7-1417-42d1-a150-42f646d04cbc {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.443412] env[63088]: DEBUG oslo_vmware.api [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Waiting for the task: (returnval){ [ 965.443412] env[63088]: value = "task-1285316" [ 965.443412] env[63088]: _type = "Task" [ 965.443412] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.448229] env[63088]: DEBUG nova.policy [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '139b65d9cf314b3db1d685d63bc19bba', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e48f62dee9ad4e9b94b67bd871db5c63', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 965.457733] env[63088]: DEBUG oslo_vmware.api [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285316, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.556050] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52bcacc8-4e7f-405a-978d-d689ac13709a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.564840] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f558a36b-c4fa-4e0d-828d-88d873ffd6f0 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.600231] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9044d46f-6a47-46bb-ae08-2fa77ab2a6fe {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.609097] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a395ce4c-7bcc-4245-ac6c-c998e14132e0 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.623079] env[63088]: DEBUG nova.compute.provider_tree [None req-b5ff0ab0-0a14-4514-b203-4512f0d101da tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 965.777812] env[63088]: DEBUG nova.network.neutron [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2ad33254-8030-4454-b023-3359071f85de] Updating instance_info_cache with network_info: [{"id": "1d2bee7e-bbc7-4b17-976b-47e7bbb268af", "address": "fa:16:3e:34:9a:ff", "network": {"id": "dff14a7f-0af1-4e4e-a498-86d7c9816e6b", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-762073671-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f33f2701fad94864a8c406a404bc0a42", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "21310d90-efbc-45a8-a97f-c4358606530f", "external-id": "nsx-vlan-transportzone-672", "segmentation_id": 672, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d2bee7e-bb", "ovs_interfaceid": "1d2bee7e-bbc7-4b17-976b-47e7bbb268af", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 965.869336] env[63088]: DEBUG nova.network.neutron [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 66178bd0-7ddf-458c-98a2-3519e536dc63] Successfully created port: 725d6827-6db3-4f39-8374-df9d333b77bb {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 965.880490] env[63088]: DEBUG nova.compute.manager [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 66178bd0-7ddf-458c-98a2-3519e536dc63] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 965.913689] env[63088]: DEBUG oslo_concurrency.lockutils [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Acquiring lock "refresh_cache-818704d4-2010-4492-8e51-28ad0bd6f570" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 965.913842] env[63088]: DEBUG oslo_concurrency.lockutils [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Acquired lock "refresh_cache-818704d4-2010-4492-8e51-28ad0bd6f570" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 965.914288] env[63088]: DEBUG nova.network.neutron [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 818704d4-2010-4492-8e51-28ad0bd6f570] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 965.954085] env[63088]: DEBUG oslo_vmware.api [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285316, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.146082} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.954350] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 965.954568] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Deleted contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 965.954767] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 965.973944] env[63088]: INFO nova.scheduler.client.report [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Deleted allocations for instance 71e37d8e-a454-46c4-a3cc-3d5671a32beb [ 966.126679] env[63088]: DEBUG nova.scheduler.client.report [None req-b5ff0ab0-0a14-4514-b203-4512f0d101da tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 966.283719] env[63088]: DEBUG oslo_concurrency.lockutils [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Releasing lock "refresh_cache-2ad33254-8030-4454-b023-3359071f85de" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 966.446969] env[63088]: DEBUG nova.network.neutron [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 818704d4-2010-4492-8e51-28ad0bd6f570] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 966.478855] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 966.499814] env[63088]: DEBUG nova.compute.manager [req-06219a0a-c007-427d-a2fe-69e1e5a89892 req-08062a8e-5e97-433f-8e75-3a9e803cae7e service nova] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Received event network-vif-unplugged-79ca8279-a3e2-4792-a4b5-57f899be5c41 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 966.500051] env[63088]: DEBUG oslo_concurrency.lockutils [req-06219a0a-c007-427d-a2fe-69e1e5a89892 req-08062a8e-5e97-433f-8e75-3a9e803cae7e service nova] Acquiring lock "71e37d8e-a454-46c4-a3cc-3d5671a32beb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 966.500297] env[63088]: DEBUG oslo_concurrency.lockutils [req-06219a0a-c007-427d-a2fe-69e1e5a89892 req-08062a8e-5e97-433f-8e75-3a9e803cae7e service nova] Lock "71e37d8e-a454-46c4-a3cc-3d5671a32beb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.500483] env[63088]: DEBUG oslo_concurrency.lockutils [req-06219a0a-c007-427d-a2fe-69e1e5a89892 req-08062a8e-5e97-433f-8e75-3a9e803cae7e service nova] Lock "71e37d8e-a454-46c4-a3cc-3d5671a32beb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.500658] env[63088]: DEBUG nova.compute.manager [req-06219a0a-c007-427d-a2fe-69e1e5a89892 req-08062a8e-5e97-433f-8e75-3a9e803cae7e service nova] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] No waiting events found dispatching network-vif-unplugged-79ca8279-a3e2-4792-a4b5-57f899be5c41 {{(pid=63088) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 966.500833] env[63088]: WARNING nova.compute.manager [req-06219a0a-c007-427d-a2fe-69e1e5a89892 req-08062a8e-5e97-433f-8e75-3a9e803cae7e service nova] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Received unexpected event network-vif-unplugged-79ca8279-a3e2-4792-a4b5-57f899be5c41 for instance with vm_state shelved_offloaded and task_state None. [ 966.501011] env[63088]: DEBUG nova.compute.manager [req-06219a0a-c007-427d-a2fe-69e1e5a89892 req-08062a8e-5e97-433f-8e75-3a9e803cae7e service nova] [instance: 818704d4-2010-4492-8e51-28ad0bd6f570] Received event network-vif-plugged-2a414d49-0028-4cbf-bc84-04946140a76c {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 966.501197] env[63088]: DEBUG oslo_concurrency.lockutils [req-06219a0a-c007-427d-a2fe-69e1e5a89892 req-08062a8e-5e97-433f-8e75-3a9e803cae7e service nova] Acquiring lock "818704d4-2010-4492-8e51-28ad0bd6f570-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 966.501415] env[63088]: DEBUG oslo_concurrency.lockutils [req-06219a0a-c007-427d-a2fe-69e1e5a89892 req-08062a8e-5e97-433f-8e75-3a9e803cae7e service nova] Lock "818704d4-2010-4492-8e51-28ad0bd6f570-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.501589] env[63088]: DEBUG oslo_concurrency.lockutils [req-06219a0a-c007-427d-a2fe-69e1e5a89892 req-08062a8e-5e97-433f-8e75-3a9e803cae7e service nova] Lock "818704d4-2010-4492-8e51-28ad0bd6f570-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.501760] env[63088]: DEBUG nova.compute.manager [req-06219a0a-c007-427d-a2fe-69e1e5a89892 req-08062a8e-5e97-433f-8e75-3a9e803cae7e service nova] [instance: 818704d4-2010-4492-8e51-28ad0bd6f570] No waiting events found dispatching network-vif-plugged-2a414d49-0028-4cbf-bc84-04946140a76c {{(pid=63088) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 966.501967] env[63088]: WARNING nova.compute.manager [req-06219a0a-c007-427d-a2fe-69e1e5a89892 req-08062a8e-5e97-433f-8e75-3a9e803cae7e service nova] [instance: 818704d4-2010-4492-8e51-28ad0bd6f570] Received unexpected event network-vif-plugged-2a414d49-0028-4cbf-bc84-04946140a76c for instance with vm_state building and task_state spawning. [ 966.502130] env[63088]: DEBUG nova.compute.manager [req-06219a0a-c007-427d-a2fe-69e1e5a89892 req-08062a8e-5e97-433f-8e75-3a9e803cae7e service nova] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Received event network-changed-79ca8279-a3e2-4792-a4b5-57f899be5c41 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 966.502276] env[63088]: DEBUG nova.compute.manager [req-06219a0a-c007-427d-a2fe-69e1e5a89892 req-08062a8e-5e97-433f-8e75-3a9e803cae7e service nova] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Refreshing instance network info cache due to event network-changed-79ca8279-a3e2-4792-a4b5-57f899be5c41. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 966.502488] env[63088]: DEBUG oslo_concurrency.lockutils [req-06219a0a-c007-427d-a2fe-69e1e5a89892 req-08062a8e-5e97-433f-8e75-3a9e803cae7e service nova] Acquiring lock "refresh_cache-71e37d8e-a454-46c4-a3cc-3d5671a32beb" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 966.502643] env[63088]: DEBUG oslo_concurrency.lockutils [req-06219a0a-c007-427d-a2fe-69e1e5a89892 req-08062a8e-5e97-433f-8e75-3a9e803cae7e service nova] Acquired lock "refresh_cache-71e37d8e-a454-46c4-a3cc-3d5671a32beb" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 966.502780] env[63088]: DEBUG nova.network.neutron [req-06219a0a-c007-427d-a2fe-69e1e5a89892 req-08062a8e-5e97-433f-8e75-3a9e803cae7e service nova] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Refreshing network info cache for port 79ca8279-a3e2-4792-a4b5-57f899be5c41 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 966.629470] env[63088]: DEBUG nova.network.neutron [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 818704d4-2010-4492-8e51-28ad0bd6f570] Updating instance_info_cache with network_info: [{"id": "2a414d49-0028-4cbf-bc84-04946140a76c", "address": "fa:16:3e:67:f7:fa", "network": {"id": "2bbaf351-fe63-46ba-ba7e-0f432c0a1bfb", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1438884226-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7bff1382b9694df08133c88a5fe783a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cca1f087-01e1-49ca-831b-5c51478a5d60", "external-id": "nsx-vlan-transportzone-439", "segmentation_id": 439, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2a414d49-00", "ovs_interfaceid": "2a414d49-0028-4cbf-bc84-04946140a76c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 966.633078] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b5ff0ab0-0a14-4514-b203-4512f0d101da tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.761s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.634959] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d3109219-0d50-4f6b-8903-47cb3c3c2bc0 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.036s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.635323] env[63088]: DEBUG nova.objects.instance [None req-d3109219-0d50-4f6b-8903-47cb3c3c2bc0 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Lazy-loading 'resources' on Instance uuid 4a56d74c-cb1f-4edf-99f0-4b54bf357a3a {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 966.652685] env[63088]: INFO nova.scheduler.client.report [None req-b5ff0ab0-0a14-4514-b203-4512f0d101da tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Deleted allocations for instance 254db932-35f1-42e3-9207-cd886efd65a3 [ 966.808439] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb005ae7-41fc-49db-88fa-abdd8034f97b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.828917] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50c3b053-a4c3-474b-9061-169f3e0ef022 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.836675] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2ad33254-8030-4454-b023-3359071f85de] Updating instance '2ad33254-8030-4454-b023-3359071f85de' progress to 83 {{(pid=63088) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 966.891955] env[63088]: DEBUG nova.compute.manager [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 66178bd0-7ddf-458c-98a2-3519e536dc63] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 966.917081] env[63088]: DEBUG nova.virt.hardware [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 966.917081] env[63088]: DEBUG nova.virt.hardware [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 966.917081] env[63088]: DEBUG nova.virt.hardware [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 966.917081] env[63088]: DEBUG nova.virt.hardware [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 966.917523] env[63088]: DEBUG nova.virt.hardware [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 966.917849] env[63088]: DEBUG nova.virt.hardware [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 966.918261] env[63088]: DEBUG nova.virt.hardware [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 966.918559] env[63088]: DEBUG nova.virt.hardware [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 966.918852] env[63088]: DEBUG nova.virt.hardware [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 966.919177] env[63088]: DEBUG nova.virt.hardware [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 966.919483] env[63088]: DEBUG nova.virt.hardware [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 966.920807] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7c21d8e-0ba2-453d-aae6-f93b9e92813f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.930332] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdc01f65-5335-4677-9219-74351441f9c6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.131686] env[63088]: DEBUG oslo_concurrency.lockutils [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Releasing lock "refresh_cache-818704d4-2010-4492-8e51-28ad0bd6f570" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 967.132049] env[63088]: DEBUG nova.compute.manager [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 818704d4-2010-4492-8e51-28ad0bd6f570] Instance network_info: |[{"id": "2a414d49-0028-4cbf-bc84-04946140a76c", "address": "fa:16:3e:67:f7:fa", "network": {"id": "2bbaf351-fe63-46ba-ba7e-0f432c0a1bfb", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1438884226-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7bff1382b9694df08133c88a5fe783a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cca1f087-01e1-49ca-831b-5c51478a5d60", "external-id": "nsx-vlan-transportzone-439", "segmentation_id": 439, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2a414d49-00", "ovs_interfaceid": "2a414d49-0028-4cbf-bc84-04946140a76c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 967.132509] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 818704d4-2010-4492-8e51-28ad0bd6f570] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:67:f7:fa', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cca1f087-01e1-49ca-831b-5c51478a5d60', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2a414d49-0028-4cbf-bc84-04946140a76c', 'vif_model': 'vmxnet3'}] {{(pid=63088) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 967.140534] env[63088]: DEBUG oslo.service.loopingcall [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 967.144226] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 818704d4-2010-4492-8e51-28ad0bd6f570] Creating VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 967.146704] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-13d21754-a144-43b5-84d0-8eaa53c669f5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.167029] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b5ff0ab0-0a14-4514-b203-4512f0d101da tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Lock "254db932-35f1-42e3-9207-cd886efd65a3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.648s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 967.174871] env[63088]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 967.174871] env[63088]: value = "task-1285317" [ 967.174871] env[63088]: _type = "Task" [ 967.174871] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.185643] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285317, 'name': CreateVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.288015] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e8fc3bb-0f8f-453b-a337-e55868613d26 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.296853] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4faf710c-3897-4d65-94a5-58b0f1348c70 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.329379] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a984cd0-89a2-4564-9eae-0c7b1fedc6b1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.337927] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64734c86-4c9c-4592-a88f-174c0074dbad {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.343709] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2ad33254-8030-4454-b023-3359071f85de] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 967.343992] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1aa1e6c2-35ea-4e67-9bcd-05a56914607c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.356189] env[63088]: DEBUG oslo_concurrency.lockutils [None req-47d36dda-27bb-484d-9238-4b4207ac4046 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquiring lock "interface-f582da15-dfc3-45e1-a995-9dd0c9533869-051455f3-4c94-4050-8610-bd8ace9f3988" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 967.356454] env[63088]: DEBUG oslo_concurrency.lockutils [None req-47d36dda-27bb-484d-9238-4b4207ac4046 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Lock "interface-f582da15-dfc3-45e1-a995-9dd0c9533869-051455f3-4c94-4050-8610-bd8ace9f3988" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 967.356808] env[63088]: DEBUG nova.objects.instance [None req-47d36dda-27bb-484d-9238-4b4207ac4046 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Lazy-loading 'flavor' on Instance uuid f582da15-dfc3-45e1-a995-9dd0c9533869 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 967.358442] env[63088]: DEBUG nova.compute.provider_tree [None req-d3109219-0d50-4f6b-8903-47cb3c3c2bc0 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 967.362714] env[63088]: DEBUG oslo_vmware.api [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Waiting for the task: (returnval){ [ 967.362714] env[63088]: value = "task-1285318" [ 967.362714] env[63088]: _type = "Task" [ 967.362714] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.379976] env[63088]: DEBUG oslo_vmware.api [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285318, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.464102] env[63088]: DEBUG nova.network.neutron [req-06219a0a-c007-427d-a2fe-69e1e5a89892 req-08062a8e-5e97-433f-8e75-3a9e803cae7e service nova] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Updated VIF entry in instance network info cache for port 79ca8279-a3e2-4792-a4b5-57f899be5c41. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 967.464490] env[63088]: DEBUG nova.network.neutron [req-06219a0a-c007-427d-a2fe-69e1e5a89892 req-08062a8e-5e97-433f-8e75-3a9e803cae7e service nova] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Updating instance_info_cache with network_info: [{"id": "79ca8279-a3e2-4792-a4b5-57f899be5c41", "address": "fa:16:3e:53:f8:3c", "network": {"id": "cc2c0304-8906-443a-a76e-e7a32158346f", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-1867758042-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.139", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b18ddddf2314d83addf550b8cb91977", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap79ca8279-a3", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 967.681559] env[63088]: DEBUG nova.network.neutron [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 66178bd0-7ddf-458c-98a2-3519e536dc63] Successfully updated port: 725d6827-6db3-4f39-8374-df9d333b77bb {{(pid=63088) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 967.690133] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285317, 'name': CreateVM_Task, 'duration_secs': 0.328895} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.690918] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 818704d4-2010-4492-8e51-28ad0bd6f570] Created VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 967.691963] env[63088]: DEBUG oslo_concurrency.lockutils [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 967.692071] env[63088]: DEBUG oslo_concurrency.lockutils [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 967.692390] env[63088]: DEBUG oslo_concurrency.lockutils [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 967.692943] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-38cffd39-8997-4ba5-bbb4-ee4dc520ec2d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.699302] env[63088]: DEBUG oslo_vmware.api [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Waiting for the task: (returnval){ [ 967.699302] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52d8304c-4718-2fe3-6b52-32ac19ab0a4e" [ 967.699302] env[63088]: _type = "Task" [ 967.699302] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.708655] env[63088]: DEBUG oslo_vmware.api [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52d8304c-4718-2fe3-6b52-32ac19ab0a4e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.872940] env[63088]: DEBUG nova.scheduler.client.report [None req-d3109219-0d50-4f6b-8903-47cb3c3c2bc0 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 967.887326] env[63088]: DEBUG oslo_vmware.api [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285318, 'name': PowerOnVM_Task, 'duration_secs': 0.419344} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.887655] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2ad33254-8030-4454-b023-3359071f85de] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 967.887848] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-572c4b1a-4de2-4fc8-a401-8011210c256c tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2ad33254-8030-4454-b023-3359071f85de] Updating instance '2ad33254-8030-4454-b023-3359071f85de' progress to 100 {{(pid=63088) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 967.969503] env[63088]: DEBUG oslo_concurrency.lockutils [req-06219a0a-c007-427d-a2fe-69e1e5a89892 req-08062a8e-5e97-433f-8e75-3a9e803cae7e service nova] Releasing lock "refresh_cache-71e37d8e-a454-46c4-a3cc-3d5671a32beb" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 967.969503] env[63088]: DEBUG nova.compute.manager [req-06219a0a-c007-427d-a2fe-69e1e5a89892 req-08062a8e-5e97-433f-8e75-3a9e803cae7e service nova] [instance: 818704d4-2010-4492-8e51-28ad0bd6f570] Received event network-changed-2a414d49-0028-4cbf-bc84-04946140a76c {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 967.969503] env[63088]: DEBUG nova.compute.manager [req-06219a0a-c007-427d-a2fe-69e1e5a89892 req-08062a8e-5e97-433f-8e75-3a9e803cae7e service nova] [instance: 818704d4-2010-4492-8e51-28ad0bd6f570] Refreshing instance network info cache due to event network-changed-2a414d49-0028-4cbf-bc84-04946140a76c. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 967.969503] env[63088]: DEBUG oslo_concurrency.lockutils [req-06219a0a-c007-427d-a2fe-69e1e5a89892 req-08062a8e-5e97-433f-8e75-3a9e803cae7e service nova] Acquiring lock "refresh_cache-818704d4-2010-4492-8e51-28ad0bd6f570" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 967.969503] env[63088]: DEBUG oslo_concurrency.lockutils [req-06219a0a-c007-427d-a2fe-69e1e5a89892 req-08062a8e-5e97-433f-8e75-3a9e803cae7e service nova] Acquired lock "refresh_cache-818704d4-2010-4492-8e51-28ad0bd6f570" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 967.969503] env[63088]: DEBUG nova.network.neutron [req-06219a0a-c007-427d-a2fe-69e1e5a89892 req-08062a8e-5e97-433f-8e75-3a9e803cae7e service nova] [instance: 818704d4-2010-4492-8e51-28ad0bd6f570] Refreshing network info cache for port 2a414d49-0028-4cbf-bc84-04946140a76c {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 967.985728] env[63088]: DEBUG nova.objects.instance [None req-47d36dda-27bb-484d-9238-4b4207ac4046 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Lazy-loading 'pci_requests' on Instance uuid f582da15-dfc3-45e1-a995-9dd0c9533869 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 968.184556] env[63088]: DEBUG oslo_concurrency.lockutils [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquiring lock "refresh_cache-66178bd0-7ddf-458c-98a2-3519e536dc63" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 968.184720] env[63088]: DEBUG oslo_concurrency.lockutils [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquired lock "refresh_cache-66178bd0-7ddf-458c-98a2-3519e536dc63" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 968.184901] env[63088]: DEBUG nova.network.neutron [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 66178bd0-7ddf-458c-98a2-3519e536dc63] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 968.210943] env[63088]: DEBUG oslo_vmware.api [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52d8304c-4718-2fe3-6b52-32ac19ab0a4e, 'name': SearchDatastore_Task, 'duration_secs': 0.011196} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.211326] env[63088]: DEBUG oslo_concurrency.lockutils [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 968.211567] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 818704d4-2010-4492-8e51-28ad0bd6f570] Processing image 1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 968.211800] env[63088]: DEBUG oslo_concurrency.lockutils [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 968.211950] env[63088]: DEBUG oslo_concurrency.lockutils [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 968.212147] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 968.212460] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ffba88a3-18f8-44a1-b450-990c41038032 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.221441] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 968.221657] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 968.222508] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ea75710c-62e5-4912-8181-a25f804cc42d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.228020] env[63088]: DEBUG oslo_vmware.api [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Waiting for the task: (returnval){ [ 968.228020] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52406d19-3c57-bb22-20ff-7052c12286a6" [ 968.228020] env[63088]: _type = "Task" [ 968.228020] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.235807] env[63088]: DEBUG oslo_vmware.api [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52406d19-3c57-bb22-20ff-7052c12286a6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.381556] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d3109219-0d50-4f6b-8903-47cb3c3c2bc0 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.746s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.383925] env[63088]: DEBUG oslo_concurrency.lockutils [None req-093bbf73-8b98-46e9-b1d0-a1fd756e135a tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.189s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.384182] env[63088]: DEBUG nova.objects.instance [None req-093bbf73-8b98-46e9-b1d0-a1fd756e135a tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lazy-loading 'resources' on Instance uuid fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 968.403069] env[63088]: INFO nova.scheduler.client.report [None req-d3109219-0d50-4f6b-8903-47cb3c3c2bc0 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Deleted allocations for instance 4a56d74c-cb1f-4edf-99f0-4b54bf357a3a [ 968.488781] env[63088]: DEBUG nova.objects.base [None req-47d36dda-27bb-484d-9238-4b4207ac4046 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=63088) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 968.488781] env[63088]: DEBUG nova.network.neutron [None req-47d36dda-27bb-484d-9238-4b4207ac4046 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 968.528205] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquiring lock "69645f50-46fc-4c15-9b39-1721f7636e31" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 968.528465] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lock "69645f50-46fc-4c15-9b39-1721f7636e31" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.647422] env[63088]: DEBUG nova.policy [None req-47d36dda-27bb-484d-9238-4b4207ac4046 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '93f1d334e71a4a99ba300c3ee6d70243', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'aa3d24a1a6c0430985fd80365d986ee1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 968.694201] env[63088]: DEBUG nova.compute.manager [req-a7685a48-8e44-4e68-9d60-51594216e41e req-8878c60f-082f-4340-8cda-c0947d04674a service nova] [instance: 66178bd0-7ddf-458c-98a2-3519e536dc63] Received event network-vif-plugged-725d6827-6db3-4f39-8374-df9d333b77bb {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 968.694441] env[63088]: DEBUG oslo_concurrency.lockutils [req-a7685a48-8e44-4e68-9d60-51594216e41e req-8878c60f-082f-4340-8cda-c0947d04674a service nova] Acquiring lock "66178bd0-7ddf-458c-98a2-3519e536dc63-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 968.696969] env[63088]: DEBUG oslo_concurrency.lockutils [req-a7685a48-8e44-4e68-9d60-51594216e41e req-8878c60f-082f-4340-8cda-c0947d04674a service nova] Lock "66178bd0-7ddf-458c-98a2-3519e536dc63-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.697208] env[63088]: DEBUG oslo_concurrency.lockutils [req-a7685a48-8e44-4e68-9d60-51594216e41e req-8878c60f-082f-4340-8cda-c0947d04674a service nova] Lock "66178bd0-7ddf-458c-98a2-3519e536dc63-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.003s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.697482] env[63088]: DEBUG nova.compute.manager [req-a7685a48-8e44-4e68-9d60-51594216e41e req-8878c60f-082f-4340-8cda-c0947d04674a service nova] [instance: 66178bd0-7ddf-458c-98a2-3519e536dc63] No waiting events found dispatching network-vif-plugged-725d6827-6db3-4f39-8374-df9d333b77bb {{(pid=63088) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 968.697692] env[63088]: WARNING nova.compute.manager [req-a7685a48-8e44-4e68-9d60-51594216e41e req-8878c60f-082f-4340-8cda-c0947d04674a service nova] [instance: 66178bd0-7ddf-458c-98a2-3519e536dc63] Received unexpected event network-vif-plugged-725d6827-6db3-4f39-8374-df9d333b77bb for instance with vm_state building and task_state spawning. [ 968.697901] env[63088]: DEBUG nova.compute.manager [req-a7685a48-8e44-4e68-9d60-51594216e41e req-8878c60f-082f-4340-8cda-c0947d04674a service nova] [instance: 66178bd0-7ddf-458c-98a2-3519e536dc63] Received event network-changed-725d6827-6db3-4f39-8374-df9d333b77bb {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 968.698070] env[63088]: DEBUG nova.compute.manager [req-a7685a48-8e44-4e68-9d60-51594216e41e req-8878c60f-082f-4340-8cda-c0947d04674a service nova] [instance: 66178bd0-7ddf-458c-98a2-3519e536dc63] Refreshing instance network info cache due to event network-changed-725d6827-6db3-4f39-8374-df9d333b77bb. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 968.698285] env[63088]: DEBUG oslo_concurrency.lockutils [req-a7685a48-8e44-4e68-9d60-51594216e41e req-8878c60f-082f-4340-8cda-c0947d04674a service nova] Acquiring lock "refresh_cache-66178bd0-7ddf-458c-98a2-3519e536dc63" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 968.724527] env[63088]: DEBUG nova.network.neutron [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 66178bd0-7ddf-458c-98a2-3519e536dc63] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 968.739978] env[63088]: DEBUG oslo_vmware.api [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52406d19-3c57-bb22-20ff-7052c12286a6, 'name': SearchDatastore_Task, 'duration_secs': 0.008744} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.748291] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2d85f4be-429a-4bd0-9f2a-71a3c5115f8d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.754112] env[63088]: DEBUG oslo_vmware.api [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Waiting for the task: (returnval){ [ 968.754112] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]529ebd62-2798-a073-c14a-fdc6461781ae" [ 968.754112] env[63088]: _type = "Task" [ 968.754112] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.762023] env[63088]: DEBUG oslo_vmware.api [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]529ebd62-2798-a073-c14a-fdc6461781ae, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.899400] env[63088]: DEBUG nova.network.neutron [req-06219a0a-c007-427d-a2fe-69e1e5a89892 req-08062a8e-5e97-433f-8e75-3a9e803cae7e service nova] [instance: 818704d4-2010-4492-8e51-28ad0bd6f570] Updated VIF entry in instance network info cache for port 2a414d49-0028-4cbf-bc84-04946140a76c. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 968.900375] env[63088]: DEBUG nova.network.neutron [req-06219a0a-c007-427d-a2fe-69e1e5a89892 req-08062a8e-5e97-433f-8e75-3a9e803cae7e service nova] [instance: 818704d4-2010-4492-8e51-28ad0bd6f570] Updating instance_info_cache with network_info: [{"id": "2a414d49-0028-4cbf-bc84-04946140a76c", "address": "fa:16:3e:67:f7:fa", "network": {"id": "2bbaf351-fe63-46ba-ba7e-0f432c0a1bfb", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1438884226-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7bff1382b9694df08133c88a5fe783a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cca1f087-01e1-49ca-831b-5c51478a5d60", "external-id": "nsx-vlan-transportzone-439", "segmentation_id": 439, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2a414d49-00", "ovs_interfaceid": "2a414d49-0028-4cbf-bc84-04946140a76c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 968.913355] env[63088]: DEBUG oslo_concurrency.lockutils [req-06219a0a-c007-427d-a2fe-69e1e5a89892 req-08062a8e-5e97-433f-8e75-3a9e803cae7e service nova] Releasing lock "refresh_cache-818704d4-2010-4492-8e51-28ad0bd6f570" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 968.919837] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d3109219-0d50-4f6b-8903-47cb3c3c2bc0 tempest-ServersAaction247Test-1561584366 tempest-ServersAaction247Test-1561584366-project-member] Lock "4a56d74c-cb1f-4edf-99f0-4b54bf357a3a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.148s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.925358] env[63088]: DEBUG nova.network.neutron [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 66178bd0-7ddf-458c-98a2-3519e536dc63] Updating instance_info_cache with network_info: [{"id": "725d6827-6db3-4f39-8374-df9d333b77bb", "address": "fa:16:3e:a8:c7:d3", "network": {"id": "5b801381-6bb3-45cd-b81f-92b5d757e080", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1481457494-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e48f62dee9ad4e9b94b67bd871db5c63", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "051f343d-ac4f-4070-a26d-467603122c81", "external-id": "nsx-vlan-transportzone-277", "segmentation_id": 277, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap725d6827-6d", "ovs_interfaceid": "725d6827-6db3-4f39-8374-df9d333b77bb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 969.031035] env[63088]: DEBUG nova.compute.manager [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 969.088597] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f692bef6-5b1d-437f-baec-217c7aae07d8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.099283] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d44c310-f074-44b1-ad60-8f0f8701306a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.102461] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 969.102745] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 969.102930] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Starting heal instance info cache {{(pid=63088) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 969.131306] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64f06312-23d5-4cf5-ab1c-4de093086681 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.139368] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9ff0205-778d-4b50-bbbc-809620866205 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.153262] env[63088]: DEBUG nova.compute.provider_tree [None req-093bbf73-8b98-46e9-b1d0-a1fd756e135a tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 969.265362] env[63088]: DEBUG oslo_vmware.api [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]529ebd62-2798-a073-c14a-fdc6461781ae, 'name': SearchDatastore_Task, 'duration_secs': 0.00995} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.265694] env[63088]: DEBUG oslo_concurrency.lockutils [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 969.265957] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore2] 818704d4-2010-4492-8e51-28ad0bd6f570/818704d4-2010-4492-8e51-28ad0bd6f570.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 969.266248] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9a08a088-f5ca-4da3-b9e3-d44491b34298 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.274350] env[63088]: DEBUG oslo_vmware.api [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Waiting for the task: (returnval){ [ 969.274350] env[63088]: value = "task-1285319" [ 969.274350] env[63088]: _type = "Task" [ 969.274350] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.288048] env[63088]: DEBUG oslo_vmware.api [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285319, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.427097] env[63088]: DEBUG oslo_concurrency.lockutils [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Releasing lock "refresh_cache-66178bd0-7ddf-458c-98a2-3519e536dc63" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 969.427377] env[63088]: DEBUG nova.compute.manager [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 66178bd0-7ddf-458c-98a2-3519e536dc63] Instance network_info: |[{"id": "725d6827-6db3-4f39-8374-df9d333b77bb", "address": "fa:16:3e:a8:c7:d3", "network": {"id": "5b801381-6bb3-45cd-b81f-92b5d757e080", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1481457494-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e48f62dee9ad4e9b94b67bd871db5c63", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "051f343d-ac4f-4070-a26d-467603122c81", "external-id": "nsx-vlan-transportzone-277", "segmentation_id": 277, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap725d6827-6d", "ovs_interfaceid": "725d6827-6db3-4f39-8374-df9d333b77bb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 969.427681] env[63088]: DEBUG oslo_concurrency.lockutils [req-a7685a48-8e44-4e68-9d60-51594216e41e req-8878c60f-082f-4340-8cda-c0947d04674a service nova] Acquired lock "refresh_cache-66178bd0-7ddf-458c-98a2-3519e536dc63" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 969.427865] env[63088]: DEBUG nova.network.neutron [req-a7685a48-8e44-4e68-9d60-51594216e41e req-8878c60f-082f-4340-8cda-c0947d04674a service nova] [instance: 66178bd0-7ddf-458c-98a2-3519e536dc63] Refreshing network info cache for port 725d6827-6db3-4f39-8374-df9d333b77bb {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 969.428983] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 66178bd0-7ddf-458c-98a2-3519e536dc63] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a8:c7:d3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '051f343d-ac4f-4070-a26d-467603122c81', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '725d6827-6db3-4f39-8374-df9d333b77bb', 'vif_model': 'vmxnet3'}] {{(pid=63088) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 969.436513] env[63088]: DEBUG oslo.service.loopingcall [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 969.440310] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 66178bd0-7ddf-458c-98a2-3519e536dc63] Creating VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 969.440934] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f747aa98-e6f0-4f5d-8166-6889ecee94d8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.465510] env[63088]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 969.465510] env[63088]: value = "task-1285320" [ 969.465510] env[63088]: _type = "Task" [ 969.465510] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.475209] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285320, 'name': CreateVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.556539] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.656898] env[63088]: DEBUG nova.scheduler.client.report [None req-093bbf73-8b98-46e9-b1d0-a1fd756e135a tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 969.711315] env[63088]: DEBUG nova.network.neutron [req-a7685a48-8e44-4e68-9d60-51594216e41e req-8878c60f-082f-4340-8cda-c0947d04674a service nova] [instance: 66178bd0-7ddf-458c-98a2-3519e536dc63] Updated VIF entry in instance network info cache for port 725d6827-6db3-4f39-8374-df9d333b77bb. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 969.711807] env[63088]: DEBUG nova.network.neutron [req-a7685a48-8e44-4e68-9d60-51594216e41e req-8878c60f-082f-4340-8cda-c0947d04674a service nova] [instance: 66178bd0-7ddf-458c-98a2-3519e536dc63] Updating instance_info_cache with network_info: [{"id": "725d6827-6db3-4f39-8374-df9d333b77bb", "address": "fa:16:3e:a8:c7:d3", "network": {"id": "5b801381-6bb3-45cd-b81f-92b5d757e080", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1481457494-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e48f62dee9ad4e9b94b67bd871db5c63", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "051f343d-ac4f-4070-a26d-467603122c81", "external-id": "nsx-vlan-transportzone-277", "segmentation_id": 277, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap725d6827-6d", "ovs_interfaceid": "725d6827-6db3-4f39-8374-df9d333b77bb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 969.786685] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Acquiring lock "71e37d8e-a454-46c4-a3cc-3d5671a32beb" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.786993] env[63088]: DEBUG oslo_vmware.api [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285319, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.978699] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285320, 'name': CreateVM_Task, 'duration_secs': 0.401287} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.979035] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 66178bd0-7ddf-458c-98a2-3519e536dc63] Created VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 969.980091] env[63088]: DEBUG oslo_concurrency.lockutils [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 969.980395] env[63088]: DEBUG oslo_concurrency.lockutils [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 969.980844] env[63088]: DEBUG oslo_concurrency.lockutils [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 969.981236] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-320e5126-be0e-46b1-bcea-80f7a9aec3a8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.987208] env[63088]: DEBUG oslo_vmware.api [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 969.987208] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]521b740c-0543-d968-0d04-0e893cb2185d" [ 969.987208] env[63088]: _type = "Task" [ 969.987208] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.996769] env[63088]: DEBUG oslo_vmware.api [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]521b740c-0543-d968-0d04-0e893cb2185d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.047758] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bc161c1d-84e4-48a5-a6e5-b5d8ec5b62aa tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquiring lock "2ad33254-8030-4454-b023-3359071f85de" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.047988] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bc161c1d-84e4-48a5-a6e5-b5d8ec5b62aa tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Lock "2ad33254-8030-4454-b023-3359071f85de" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.048221] env[63088]: DEBUG nova.compute.manager [None req-bc161c1d-84e4-48a5-a6e5-b5d8ec5b62aa tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2ad33254-8030-4454-b023-3359071f85de] Going to confirm migration 3 {{(pid=63088) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 970.162663] env[63088]: DEBUG oslo_concurrency.lockutils [None req-093bbf73-8b98-46e9-b1d0-a1fd756e135a tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.779s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 970.165325] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a6622018-576f-4fec-b257-a0dbafc3218c tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.114s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.165597] env[63088]: DEBUG nova.objects.instance [None req-a6622018-576f-4fec-b257-a0dbafc3218c tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Lazy-loading 'resources' on Instance uuid 49ee4180-6322-4003-a5b1-f2a91d190290 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 970.188932] env[63088]: INFO nova.scheduler.client.report [None req-093bbf73-8b98-46e9-b1d0-a1fd756e135a tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Deleted allocations for instance fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3 [ 970.215017] env[63088]: DEBUG oslo_concurrency.lockutils [req-a7685a48-8e44-4e68-9d60-51594216e41e req-8878c60f-082f-4340-8cda-c0947d04674a service nova] Releasing lock "refresh_cache-66178bd0-7ddf-458c-98a2-3519e536dc63" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 970.286060] env[63088]: DEBUG oslo_vmware.api [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285319, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.566126} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.286354] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore2] 818704d4-2010-4492-8e51-28ad0bd6f570/818704d4-2010-4492-8e51-28ad0bd6f570.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 970.286656] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 818704d4-2010-4492-8e51-28ad0bd6f570] Extending root virtual disk to 1048576 {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 970.286934] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0ffa628b-0e68-44ef-992f-5a6d9fe32e51 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.295547] env[63088]: DEBUG oslo_vmware.api [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Waiting for the task: (returnval){ [ 970.295547] env[63088]: value = "task-1285321" [ 970.295547] env[63088]: _type = "Task" [ 970.295547] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.305226] env[63088]: DEBUG oslo_vmware.api [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285321, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.434312] env[63088]: DEBUG nova.network.neutron [None req-47d36dda-27bb-484d-9238-4b4207ac4046 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Successfully updated port: 051455f3-4c94-4050-8610-bd8ace9f3988 {{(pid=63088) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 970.499316] env[63088]: DEBUG oslo_vmware.api [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]521b740c-0543-d968-0d04-0e893cb2185d, 'name': SearchDatastore_Task, 'duration_secs': 0.012417} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.499607] env[63088]: DEBUG oslo_concurrency.lockutils [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 970.499852] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 66178bd0-7ddf-458c-98a2-3519e536dc63] Processing image 1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 970.500116] env[63088]: DEBUG oslo_concurrency.lockutils [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 970.500303] env[63088]: DEBUG oslo_concurrency.lockutils [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 970.500494] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 970.500813] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2da0abac-2eb5-4ba0-8b4c-d4a52072c92c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.511226] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 970.511425] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 970.512216] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f78bc90b-d92a-4c9c-a93b-a05eb5d026de {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.517751] env[63088]: DEBUG oslo_vmware.api [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 970.517751] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52edb139-313e-4f2a-9a23-361447b877a4" [ 970.517751] env[63088]: _type = "Task" [ 970.517751] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.527834] env[63088]: DEBUG oslo_vmware.api [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52edb139-313e-4f2a-9a23-361447b877a4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.586502] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bc161c1d-84e4-48a5-a6e5-b5d8ec5b62aa tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquiring lock "refresh_cache-2ad33254-8030-4454-b023-3359071f85de" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 970.586786] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bc161c1d-84e4-48a5-a6e5-b5d8ec5b62aa tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquired lock "refresh_cache-2ad33254-8030-4454-b023-3359071f85de" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 970.586938] env[63088]: DEBUG nova.network.neutron [None req-bc161c1d-84e4-48a5-a6e5-b5d8ec5b62aa tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2ad33254-8030-4454-b023-3359071f85de] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 970.587196] env[63088]: DEBUG nova.objects.instance [None req-bc161c1d-84e4-48a5-a6e5-b5d8ec5b62aa tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Lazy-loading 'info_cache' on Instance uuid 2ad33254-8030-4454-b023-3359071f85de {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 970.695695] env[63088]: DEBUG oslo_concurrency.lockutils [None req-093bbf73-8b98-46e9-b1d0-a1fd756e135a tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lock "fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.774s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 970.794991] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27f58978-ba30-4a67-97dc-1374662aa508 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.808504] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f29b7ce4-8731-445e-8fb5-b199e107abfc {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.811720] env[63088]: DEBUG oslo_vmware.api [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285321, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075751} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.811960] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 818704d4-2010-4492-8e51-28ad0bd6f570] Extended root virtual disk {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 970.813060] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-faa11a70-1655-4221-9462-2b359cad9700 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.840117] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-473a7042-f13c-499f-b814-5d95f49d08a7 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.854573] env[63088]: DEBUG nova.compute.manager [req-8de0e910-7b31-40dc-aa2f-00b6f3751ebd req-9cb45eb1-e9fb-4a72-b9da-4384581cf2b6 service nova] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Received event network-vif-plugged-051455f3-4c94-4050-8610-bd8ace9f3988 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 970.854850] env[63088]: DEBUG oslo_concurrency.lockutils [req-8de0e910-7b31-40dc-aa2f-00b6f3751ebd req-9cb45eb1-e9fb-4a72-b9da-4384581cf2b6 service nova] Acquiring lock "f582da15-dfc3-45e1-a995-9dd0c9533869-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.855095] env[63088]: DEBUG oslo_concurrency.lockutils [req-8de0e910-7b31-40dc-aa2f-00b6f3751ebd req-9cb45eb1-e9fb-4a72-b9da-4384581cf2b6 service nova] Lock "f582da15-dfc3-45e1-a995-9dd0c9533869-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.856229] env[63088]: DEBUG oslo_concurrency.lockutils [req-8de0e910-7b31-40dc-aa2f-00b6f3751ebd req-9cb45eb1-e9fb-4a72-b9da-4384581cf2b6 service nova] Lock "f582da15-dfc3-45e1-a995-9dd0c9533869-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 970.856229] env[63088]: DEBUG nova.compute.manager [req-8de0e910-7b31-40dc-aa2f-00b6f3751ebd req-9cb45eb1-e9fb-4a72-b9da-4384581cf2b6 service nova] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] No waiting events found dispatching network-vif-plugged-051455f3-4c94-4050-8610-bd8ace9f3988 {{(pid=63088) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 970.856229] env[63088]: WARNING nova.compute.manager [req-8de0e910-7b31-40dc-aa2f-00b6f3751ebd req-9cb45eb1-e9fb-4a72-b9da-4384581cf2b6 service nova] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Received unexpected event network-vif-plugged-051455f3-4c94-4050-8610-bd8ace9f3988 for instance with vm_state active and task_state None. [ 970.856229] env[63088]: DEBUG nova.compute.manager [req-8de0e910-7b31-40dc-aa2f-00b6f3751ebd req-9cb45eb1-e9fb-4a72-b9da-4384581cf2b6 service nova] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Received event network-changed-051455f3-4c94-4050-8610-bd8ace9f3988 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 970.856229] env[63088]: DEBUG nova.compute.manager [req-8de0e910-7b31-40dc-aa2f-00b6f3751ebd req-9cb45eb1-e9fb-4a72-b9da-4384581cf2b6 service nova] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Refreshing instance network info cache due to event network-changed-051455f3-4c94-4050-8610-bd8ace9f3988. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 970.856229] env[63088]: DEBUG oslo_concurrency.lockutils [req-8de0e910-7b31-40dc-aa2f-00b6f3751ebd req-9cb45eb1-e9fb-4a72-b9da-4384581cf2b6 service nova] Acquiring lock "refresh_cache-f582da15-dfc3-45e1-a995-9dd0c9533869" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 970.856492] env[63088]: DEBUG oslo_concurrency.lockutils [req-8de0e910-7b31-40dc-aa2f-00b6f3751ebd req-9cb45eb1-e9fb-4a72-b9da-4384581cf2b6 service nova] Acquired lock "refresh_cache-f582da15-dfc3-45e1-a995-9dd0c9533869" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 970.856532] env[63088]: DEBUG nova.network.neutron [req-8de0e910-7b31-40dc-aa2f-00b6f3751ebd req-9cb45eb1-e9fb-4a72-b9da-4384581cf2b6 service nova] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Refreshing network info cache for port 051455f3-4c94-4050-8610-bd8ace9f3988 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 970.866247] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 818704d4-2010-4492-8e51-28ad0bd6f570] Reconfiguring VM instance instance-0000005b to attach disk [datastore2] 818704d4-2010-4492-8e51-28ad0bd6f570/818704d4-2010-4492-8e51-28ad0bd6f570.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 970.869571] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-67945c82-ceac-4e4f-b956-506649fdaf5b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.889965] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ec54825-c1a0-4d10-85d0-ca59307c6b7d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.895707] env[63088]: DEBUG oslo_vmware.api [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Waiting for the task: (returnval){ [ 970.895707] env[63088]: value = "task-1285322" [ 970.895707] env[63088]: _type = "Task" [ 970.895707] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.907507] env[63088]: DEBUG nova.compute.provider_tree [None req-a6622018-576f-4fec-b257-a0dbafc3218c tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 970.915070] env[63088]: DEBUG oslo_vmware.api [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285322, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.937732] env[63088]: DEBUG oslo_concurrency.lockutils [None req-47d36dda-27bb-484d-9238-4b4207ac4046 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquiring lock "refresh_cache-f582da15-dfc3-45e1-a995-9dd0c9533869" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 971.028931] env[63088]: DEBUG oslo_vmware.api [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52edb139-313e-4f2a-9a23-361447b877a4, 'name': SearchDatastore_Task, 'duration_secs': 0.012633} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.029765] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-95c26914-c3af-43e9-ac49-b6d940de5a65 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.035153] env[63088]: DEBUG oslo_vmware.api [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 971.035153] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52aa8ea6-f361-9245-ca2b-f2e85a5c5911" [ 971.035153] env[63088]: _type = "Task" [ 971.035153] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.043509] env[63088]: DEBUG oslo_vmware.api [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52aa8ea6-f361-9245-ca2b-f2e85a5c5911, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.153356] env[63088]: DEBUG nova.network.neutron [req-8de0e910-7b31-40dc-aa2f-00b6f3751ebd req-9cb45eb1-e9fb-4a72-b9da-4384581cf2b6 service nova] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Added VIF to instance network info cache for port 051455f3-4c94-4050-8610-bd8ace9f3988. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3489}} [ 971.153856] env[63088]: DEBUG nova.network.neutron [req-8de0e910-7b31-40dc-aa2f-00b6f3751ebd req-9cb45eb1-e9fb-4a72-b9da-4384581cf2b6 service nova] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Updating instance_info_cache with network_info: [{"id": "81fc2dda-617e-4e4c-aa91-7e7020003995", "address": "fa:16:3e:8c:09:4d", "network": {"id": "1174a284-d3ca-4f9e-aa81-13ee9a693e55", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1994276040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.150", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa3d24a1a6c0430985fd80365d986ee1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap81fc2dda-61", "ovs_interfaceid": "81fc2dda-617e-4e4c-aa91-7e7020003995", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "051455f3-4c94-4050-8610-bd8ace9f3988", "address": "fa:16:3e:e5:97:d9", "network": {"id": "1174a284-d3ca-4f9e-aa81-13ee9a693e55", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1994276040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa3d24a1a6c0430985fd80365d986ee1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap051455f3-4c", "ovs_interfaceid": "051455f3-4c94-4050-8610-bd8ace9f3988", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 971.340155] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Acquiring lock "13388466-4adb-4d56-9fc1-e3f5dc516077" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.340424] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Lock "13388466-4adb-4d56-9fc1-e3f5dc516077" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 971.406548] env[63088]: DEBUG oslo_vmware.api [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285322, 'name': ReconfigVM_Task, 'duration_secs': 0.291524} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.406848] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 818704d4-2010-4492-8e51-28ad0bd6f570] Reconfigured VM instance instance-0000005b to attach disk [datastore2] 818704d4-2010-4492-8e51-28ad0bd6f570/818704d4-2010-4492-8e51-28ad0bd6f570.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 971.407478] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-00d7de73-80ed-4cc8-8b4b-15e610ce5a69 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.410154] env[63088]: DEBUG nova.scheduler.client.report [None req-a6622018-576f-4fec-b257-a0dbafc3218c tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 971.415648] env[63088]: DEBUG oslo_vmware.api [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Waiting for the task: (returnval){ [ 971.415648] env[63088]: value = "task-1285323" [ 971.415648] env[63088]: _type = "Task" [ 971.415648] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.424805] env[63088]: DEBUG oslo_vmware.api [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285323, 'name': Rename_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.547344] env[63088]: DEBUG oslo_vmware.api [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52aa8ea6-f361-9245-ca2b-f2e85a5c5911, 'name': SearchDatastore_Task, 'duration_secs': 0.011234} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.547643] env[63088]: DEBUG oslo_concurrency.lockutils [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 971.547900] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] 66178bd0-7ddf-458c-98a2-3519e536dc63/66178bd0-7ddf-458c-98a2-3519e536dc63.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 971.548209] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b445c9ec-17df-456b-ba99-3300b6e11105 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.555173] env[63088]: DEBUG oslo_vmware.api [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 971.555173] env[63088]: value = "task-1285324" [ 971.555173] env[63088]: _type = "Task" [ 971.555173] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.564461] env[63088]: DEBUG oslo_vmware.api [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285324, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.657694] env[63088]: DEBUG oslo_concurrency.lockutils [req-8de0e910-7b31-40dc-aa2f-00b6f3751ebd req-9cb45eb1-e9fb-4a72-b9da-4384581cf2b6 service nova] Releasing lock "refresh_cache-f582da15-dfc3-45e1-a995-9dd0c9533869" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 971.658470] env[63088]: DEBUG oslo_concurrency.lockutils [None req-47d36dda-27bb-484d-9238-4b4207ac4046 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquired lock "refresh_cache-f582da15-dfc3-45e1-a995-9dd0c9533869" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 971.658675] env[63088]: DEBUG nova.network.neutron [None req-47d36dda-27bb-484d-9238-4b4207ac4046 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 971.795097] env[63088]: DEBUG oslo_concurrency.lockutils [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Acquiring lock "c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.795343] env[63088]: DEBUG oslo_concurrency.lockutils [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Lock "c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 971.843218] env[63088]: DEBUG nova.compute.manager [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: 13388466-4adb-4d56-9fc1-e3f5dc516077] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 971.872562] env[63088]: DEBUG nova.network.neutron [None req-bc161c1d-84e4-48a5-a6e5-b5d8ec5b62aa tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2ad33254-8030-4454-b023-3359071f85de] Updating instance_info_cache with network_info: [{"id": "1d2bee7e-bbc7-4b17-976b-47e7bbb268af", "address": "fa:16:3e:34:9a:ff", "network": {"id": "dff14a7f-0af1-4e4e-a498-86d7c9816e6b", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-762073671-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f33f2701fad94864a8c406a404bc0a42", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "21310d90-efbc-45a8-a97f-c4358606530f", "external-id": "nsx-vlan-transportzone-672", "segmentation_id": 672, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d2bee7e-bb", "ovs_interfaceid": "1d2bee7e-bbc7-4b17-976b-47e7bbb268af", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 971.917415] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a6622018-576f-4fec-b257-a0dbafc3218c tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.752s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.920843] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.442s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 971.920843] env[63088]: DEBUG nova.objects.instance [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Lazy-loading 'resources' on Instance uuid 71e37d8e-a454-46c4-a3cc-3d5671a32beb {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 971.934374] env[63088]: DEBUG oslo_vmware.api [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285323, 'name': Rename_Task, 'duration_secs': 0.155512} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.934616] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 818704d4-2010-4492-8e51-28ad0bd6f570] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 971.934926] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4c8804d6-a3c6-4c23-a598-348c3b3972a1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.942443] env[63088]: INFO nova.scheduler.client.report [None req-a6622018-576f-4fec-b257-a0dbafc3218c tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Deleted allocations for instance 49ee4180-6322-4003-a5b1-f2a91d190290 [ 971.946586] env[63088]: DEBUG oslo_vmware.api [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Waiting for the task: (returnval){ [ 971.946586] env[63088]: value = "task-1285325" [ 971.946586] env[63088]: _type = "Task" [ 971.946586] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.960403] env[63088]: DEBUG oslo_vmware.api [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285325, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.065251] env[63088]: DEBUG oslo_vmware.api [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285324, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.173191] env[63088]: DEBUG oslo_concurrency.lockutils [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquiring lock "c4fb31b2-1301-4a6e-bef2-8ea462caee0d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.173438] env[63088]: DEBUG oslo_concurrency.lockutils [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lock "c4fb31b2-1301-4a6e-bef2-8ea462caee0d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 972.206142] env[63088]: WARNING nova.network.neutron [None req-47d36dda-27bb-484d-9238-4b4207ac4046 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] 1174a284-d3ca-4f9e-aa81-13ee9a693e55 already exists in list: networks containing: ['1174a284-d3ca-4f9e-aa81-13ee9a693e55']. ignoring it [ 972.206142] env[63088]: WARNING nova.network.neutron [None req-47d36dda-27bb-484d-9238-4b4207ac4046 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] 1174a284-d3ca-4f9e-aa81-13ee9a693e55 already exists in list: networks containing: ['1174a284-d3ca-4f9e-aa81-13ee9a693e55']. ignoring it [ 972.206142] env[63088]: WARNING nova.network.neutron [None req-47d36dda-27bb-484d-9238-4b4207ac4046 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] 051455f3-4c94-4050-8610-bd8ace9f3988 already exists in list: port_ids containing: ['051455f3-4c94-4050-8610-bd8ace9f3988']. ignoring it [ 972.298741] env[63088]: DEBUG nova.compute.manager [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 972.368940] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.374680] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bc161c1d-84e4-48a5-a6e5-b5d8ec5b62aa tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Releasing lock "refresh_cache-2ad33254-8030-4454-b023-3359071f85de" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 972.374950] env[63088]: DEBUG nova.objects.instance [None req-bc161c1d-84e4-48a5-a6e5-b5d8ec5b62aa tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Lazy-loading 'migration_context' on Instance uuid 2ad33254-8030-4454-b023-3359071f85de {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 972.427307] env[63088]: DEBUG nova.objects.instance [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Lazy-loading 'numa_topology' on Instance uuid 71e37d8e-a454-46c4-a3cc-3d5671a32beb {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 972.456658] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a6622018-576f-4fec-b257-a0dbafc3218c tempest-ServerMetadataNegativeTestJSON-715760941 tempest-ServerMetadataNegativeTestJSON-715760941-project-member] Lock "49ee4180-6322-4003-a5b1-f2a91d190290" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.931s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.462637] env[63088]: DEBUG oslo_vmware.api [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285325, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.565906] env[63088]: DEBUG oslo_vmware.api [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285324, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.516726} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.566126] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] 66178bd0-7ddf-458c-98a2-3519e536dc63/66178bd0-7ddf-458c-98a2-3519e536dc63.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 972.566345] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 66178bd0-7ddf-458c-98a2-3519e536dc63] Extending root virtual disk to 1048576 {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 972.566601] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0ab72b29-e05c-43a8-9cf1-3edd414b1910 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.573470] env[63088]: DEBUG oslo_vmware.api [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 972.573470] env[63088]: value = "task-1285326" [ 972.573470] env[63088]: _type = "Task" [ 972.573470] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.578210] env[63088]: DEBUG nova.network.neutron [None req-47d36dda-27bb-484d-9238-4b4207ac4046 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Updating instance_info_cache with network_info: [{"id": "81fc2dda-617e-4e4c-aa91-7e7020003995", "address": "fa:16:3e:8c:09:4d", "network": {"id": "1174a284-d3ca-4f9e-aa81-13ee9a693e55", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1994276040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.150", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa3d24a1a6c0430985fd80365d986ee1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap81fc2dda-61", "ovs_interfaceid": "81fc2dda-617e-4e4c-aa91-7e7020003995", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "051455f3-4c94-4050-8610-bd8ace9f3988", "address": "fa:16:3e:e5:97:d9", "network": {"id": "1174a284-d3ca-4f9e-aa81-13ee9a693e55", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1994276040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa3d24a1a6c0430985fd80365d986ee1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap051455f3-4c", "ovs_interfaceid": "051455f3-4c94-4050-8610-bd8ace9f3988", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 972.582492] env[63088]: DEBUG oslo_vmware.api [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285326, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.675841] env[63088]: DEBUG nova.compute.manager [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: c4fb31b2-1301-4a6e-bef2-8ea462caee0d] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 972.820167] env[63088]: DEBUG oslo_concurrency.lockutils [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.877814] env[63088]: DEBUG nova.objects.base [None req-bc161c1d-84e4-48a5-a6e5-b5d8ec5b62aa tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Object Instance<2ad33254-8030-4454-b023-3359071f85de> lazy-loaded attributes: info_cache,migration_context {{(pid=63088) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 972.878855] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea84fabd-5f04-4acb-a206-fd0decd80bc4 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.899953] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-26c0274c-0e00-41ff-ab1b-16a38b87549d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.906322] env[63088]: DEBUG oslo_vmware.api [None req-bc161c1d-84e4-48a5-a6e5-b5d8ec5b62aa tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Waiting for the task: (returnval){ [ 972.906322] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52d04092-2e06-bf0c-3414-4d723424fd22" [ 972.906322] env[63088]: _type = "Task" [ 972.906322] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.914684] env[63088]: DEBUG oslo_vmware.api [None req-bc161c1d-84e4-48a5-a6e5-b5d8ec5b62aa tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52d04092-2e06-bf0c-3414-4d723424fd22, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.929625] env[63088]: DEBUG nova.objects.base [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Object Instance<71e37d8e-a454-46c4-a3cc-3d5671a32beb> lazy-loaded attributes: resources,numa_topology {{(pid=63088) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 972.964359] env[63088]: DEBUG oslo_vmware.api [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285325, 'name': PowerOnVM_Task, 'duration_secs': 0.689227} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.964809] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 818704d4-2010-4492-8e51-28ad0bd6f570] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 972.965085] env[63088]: INFO nova.compute.manager [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 818704d4-2010-4492-8e51-28ad0bd6f570] Took 8.35 seconds to spawn the instance on the hypervisor. [ 972.965304] env[63088]: DEBUG nova.compute.manager [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 818704d4-2010-4492-8e51-28ad0bd6f570] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 972.966677] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4828669d-2fc0-4f8e-92e2-f4f21f3e6090 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.084114] env[63088]: DEBUG oslo_concurrency.lockutils [None req-47d36dda-27bb-484d-9238-4b4207ac4046 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Releasing lock "refresh_cache-f582da15-dfc3-45e1-a995-9dd0c9533869" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 973.084747] env[63088]: DEBUG oslo_concurrency.lockutils [None req-47d36dda-27bb-484d-9238-4b4207ac4046 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquiring lock "f582da15-dfc3-45e1-a995-9dd0c9533869" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 973.084908] env[63088]: DEBUG oslo_concurrency.lockutils [None req-47d36dda-27bb-484d-9238-4b4207ac4046 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquired lock "f582da15-dfc3-45e1-a995-9dd0c9533869" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 973.085188] env[63088]: DEBUG oslo_vmware.api [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285326, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.397466} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.085887] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79340cb0-ea39-4f7d-944a-cc66d1da04f3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.088296] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 66178bd0-7ddf-458c-98a2-3519e536dc63] Extended root virtual disk {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 973.089014] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2184264-82b3-42c5-9bd7-1746ba3201e5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.115694] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 66178bd0-7ddf-458c-98a2-3519e536dc63] Reconfiguring VM instance instance-0000005c to attach disk [datastore1] 66178bd0-7ddf-458c-98a2-3519e536dc63/66178bd0-7ddf-458c-98a2-3519e536dc63.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 973.127382] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3935d154-94ec-49a7-bafd-78b58efddba8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.148016] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15a41d0a-921d-4f38-9558-429c7f75867d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.148016] env[63088]: DEBUG nova.virt.hardware [None req-47d36dda-27bb-484d-9238-4b4207ac4046 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 973.148016] env[63088]: DEBUG nova.virt.hardware [None req-47d36dda-27bb-484d-9238-4b4207ac4046 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 973.148016] env[63088]: DEBUG nova.virt.hardware [None req-47d36dda-27bb-484d-9238-4b4207ac4046 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 973.148016] env[63088]: DEBUG nova.virt.hardware [None req-47d36dda-27bb-484d-9238-4b4207ac4046 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 973.148016] env[63088]: DEBUG nova.virt.hardware [None req-47d36dda-27bb-484d-9238-4b4207ac4046 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 973.148016] env[63088]: DEBUG nova.virt.hardware [None req-47d36dda-27bb-484d-9238-4b4207ac4046 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 973.148016] env[63088]: DEBUG nova.virt.hardware [None req-47d36dda-27bb-484d-9238-4b4207ac4046 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 973.148016] env[63088]: DEBUG nova.virt.hardware [None req-47d36dda-27bb-484d-9238-4b4207ac4046 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 973.148016] env[63088]: DEBUG nova.virt.hardware [None req-47d36dda-27bb-484d-9238-4b4207ac4046 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 973.148016] env[63088]: DEBUG nova.virt.hardware [None req-47d36dda-27bb-484d-9238-4b4207ac4046 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 973.148016] env[63088]: DEBUG nova.virt.hardware [None req-47d36dda-27bb-484d-9238-4b4207ac4046 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 973.152907] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-47d36dda-27bb-484d-9238-4b4207ac4046 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Reconfiguring VM to attach interface {{(pid=63088) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 973.154083] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-74a6e06a-1ef2-4354-88a3-92c017973e08 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.178189] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6aa66cf7-cba3-43cf-84cf-4e93b499371c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.184384] env[63088]: DEBUG oslo_vmware.api [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 973.184384] env[63088]: value = "task-1285327" [ 973.184384] env[63088]: _type = "Task" [ 973.184384] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.189040] env[63088]: DEBUG oslo_vmware.api [None req-47d36dda-27bb-484d-9238-4b4207ac4046 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Waiting for the task: (returnval){ [ 973.189040] env[63088]: value = "task-1285328" [ 973.189040] env[63088]: _type = "Task" [ 973.189040] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.222411] env[63088]: DEBUG oslo_concurrency.lockutils [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 973.223765] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bbc7298-1065-4594-a677-750e8958de80 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.232844] env[63088]: DEBUG oslo_vmware.api [None req-47d36dda-27bb-484d-9238-4b4207ac4046 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285328, 'name': ReconfigVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.233147] env[63088]: DEBUG oslo_vmware.api [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285327, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.238893] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-030e5e58-5d50-4d82-895a-6ac191ee0768 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.255679] env[63088]: DEBUG nova.compute.provider_tree [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 973.416785] env[63088]: DEBUG oslo_vmware.api [None req-bc161c1d-84e4-48a5-a6e5-b5d8ec5b62aa tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52d04092-2e06-bf0c-3414-4d723424fd22, 'name': SearchDatastore_Task, 'duration_secs': 0.008739} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.417123] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bc161c1d-84e4-48a5-a6e5-b5d8ec5b62aa tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 973.489797] env[63088]: INFO nova.compute.manager [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 818704d4-2010-4492-8e51-28ad0bd6f570] Took 14.25 seconds to build instance. [ 973.697254] env[63088]: DEBUG oslo_vmware.api [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285327, 'name': ReconfigVM_Task, 'duration_secs': 0.338013} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.697946] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 66178bd0-7ddf-458c-98a2-3519e536dc63] Reconfigured VM instance instance-0000005c to attach disk [datastore1] 66178bd0-7ddf-458c-98a2-3519e536dc63/66178bd0-7ddf-458c-98a2-3519e536dc63.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 973.698610] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0b00ef4d-31e6-40fc-8002-420cd328b0fc {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.703106] env[63088]: DEBUG oslo_vmware.api [None req-47d36dda-27bb-484d-9238-4b4207ac4046 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285328, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.708975] env[63088]: DEBUG oslo_vmware.api [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 973.708975] env[63088]: value = "task-1285329" [ 973.708975] env[63088]: _type = "Task" [ 973.708975] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.717358] env[63088]: DEBUG oslo_vmware.api [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285329, 'name': Rename_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.720890] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Acquiring lock "refresh_cache-7b6aadb7-e34b-42b7-b69f-370434f5b665" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 973.721027] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Acquired lock "refresh_cache-7b6aadb7-e34b-42b7-b69f-370434f5b665" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 973.721172] env[63088]: DEBUG nova.network.neutron [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Forcefully refreshing network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 973.758434] env[63088]: DEBUG nova.scheduler.client.report [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 973.993421] env[63088]: DEBUG oslo_concurrency.lockutils [None req-cb5eb443-f32d-46e9-aca0-1c5ca1032ef4 tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Lock "818704d4-2010-4492-8e51-28ad0bd6f570" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.774s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 974.203762] env[63088]: DEBUG oslo_vmware.api [None req-47d36dda-27bb-484d-9238-4b4207ac4046 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285328, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.228029] env[63088]: DEBUG oslo_vmware.api [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285329, 'name': Rename_Task, 'duration_secs': 0.142937} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.228850] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 66178bd0-7ddf-458c-98a2-3519e536dc63] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 974.229612] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e1480f88-b8bc-496b-808e-b14311cab981 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.239912] env[63088]: DEBUG oslo_vmware.api [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 974.239912] env[63088]: value = "task-1285330" [ 974.239912] env[63088]: _type = "Task" [ 974.239912] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.249274] env[63088]: DEBUG oslo_vmware.api [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285330, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.263396] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.343s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 974.266976] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.711s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 974.268581] env[63088]: INFO nova.compute.claims [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 974.707222] env[63088]: DEBUG oslo_vmware.api [None req-47d36dda-27bb-484d-9238-4b4207ac4046 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285328, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.752281] env[63088]: DEBUG oslo_vmware.api [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285330, 'name': PowerOnVM_Task} progress is 76%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.774738] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c98b84f1-ee5d-494b-b628-0d6e82fa9db6 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Lock "71e37d8e-a454-46c4-a3cc-3d5671a32beb" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 28.224s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 974.776392] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Lock "71e37d8e-a454-46c4-a3cc-3d5671a32beb" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 4.990s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 974.776622] env[63088]: INFO nova.compute.manager [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Unshelving [ 975.165994] env[63088]: DEBUG nova.network.neutron [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Updating instance_info_cache with network_info: [{"id": "fa4a91d1-df9a-4789-bc5c-a8b95457cd93", "address": "fa:16:3e:7e:c2:c7", "network": {"id": "dfcbe153-bc01-4362-9247-de9f0b1c847b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-841757397-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.152", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dace8b5181b84623b08f903d12dfd31e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c68b7663-4f0e-47f0-ac7f-40c6d952f7bb", "external-id": "nsx-vlan-transportzone-696", "segmentation_id": 696, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfa4a91d1-df", "ovs_interfaceid": "fa4a91d1-df9a-4789-bc5c-a8b95457cd93", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 975.206610] env[63088]: DEBUG oslo_vmware.api [None req-47d36dda-27bb-484d-9238-4b4207ac4046 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285328, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.251765] env[63088]: DEBUG oslo_vmware.api [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285330, 'name': PowerOnVM_Task, 'duration_secs': 0.823038} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.252092] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 66178bd0-7ddf-458c-98a2-3519e536dc63] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 975.252312] env[63088]: INFO nova.compute.manager [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 66178bd0-7ddf-458c-98a2-3519e536dc63] Took 8.36 seconds to spawn the instance on the hypervisor. [ 975.252500] env[63088]: DEBUG nova.compute.manager [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 66178bd0-7ddf-458c-98a2-3519e536dc63] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 975.253374] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c0f6f5b-8d69-4583-89fb-2cabe27b45a3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.426573] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9fb6749-eb44-4f75-8bf0-5faf9eacb542 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.434745] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8a57ac9-d8c4-408d-8222-1c3545a65429 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.464163] env[63088]: DEBUG nova.compute.manager [None req-79c1771e-6b38-47c6-b2c7-527b474cbc1c tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 818704d4-2010-4492-8e51-28ad0bd6f570] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 975.465424] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb85de70-42d2-48a0-b084-3c2426554339 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.468302] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74508712-ebba-4c34-9c3c-ba8ad2af580c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.478308] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-204a200f-c40f-4f2e-847d-f8f30a9f2a20 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.493750] env[63088]: DEBUG nova.compute.provider_tree [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 975.668890] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Releasing lock "refresh_cache-7b6aadb7-e34b-42b7-b69f-370434f5b665" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 975.669137] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Updated the network info_cache for instance {{(pid=63088) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 975.669370] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 975.669650] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 975.669852] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 975.670143] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 975.670488] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 975.670645] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 975.670775] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63088) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 975.670924] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager.update_available_resource {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 975.706306] env[63088]: DEBUG oslo_vmware.api [None req-47d36dda-27bb-484d-9238-4b4207ac4046 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285328, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.784968] env[63088]: INFO nova.compute.manager [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 66178bd0-7ddf-458c-98a2-3519e536dc63] Took 15.39 seconds to build instance. [ 975.787334] env[63088]: DEBUG nova.compute.utils [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 975.986066] env[63088]: INFO nova.compute.manager [None req-79c1771e-6b38-47c6-b2c7-527b474cbc1c tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 818704d4-2010-4492-8e51-28ad0bd6f570] instance snapshotting [ 975.988597] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d8218e2-b658-4dfc-ad94-9a19c8943408 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.008034] env[63088]: DEBUG nova.scheduler.client.report [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 976.011954] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5865bea-4608-44e4-86fb-73541312d690 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.173766] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.206730] env[63088]: DEBUG oslo_vmware.api [None req-47d36dda-27bb-484d-9238-4b4207ac4046 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285328, 'name': ReconfigVM_Task, 'duration_secs': 2.740138} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.207329] env[63088]: DEBUG oslo_concurrency.lockutils [None req-47d36dda-27bb-484d-9238-4b4207ac4046 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Releasing lock "f582da15-dfc3-45e1-a995-9dd0c9533869" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 976.207552] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-47d36dda-27bb-484d-9238-4b4207ac4046 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Reconfigured VM to attach interface {{(pid=63088) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 976.287328] env[63088]: DEBUG oslo_concurrency.lockutils [None req-7649d125-81c2-460b-98a2-7694ec1edd4e tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "66178bd0-7ddf-458c-98a2-3519e536dc63" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.897s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.290568] env[63088]: INFO nova.virt.block_device [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Booting with volume 8a71432a-16cf-4ea7-a27f-e1899ca9f9e8 at /dev/sdb [ 976.324516] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5423c49e-05c7-4745-ad37-12f10bedc544 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.334902] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-babaf698-1dc0-4e4b-a2d7-012e448f80b6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.364028] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-34245ee1-5fe8-46d2-b43d-021e99b8031d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.373162] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9465d62-5d2f-488b-8c60-2665afd77e25 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.400906] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54a140aa-8cbf-45c9-8ceb-bb89883be2e2 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.408122] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7393c7b-a04c-47df-8e1b-96a673697f7c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.423081] env[63088]: DEBUG nova.virt.block_device [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Updating existing volume attachment record: 8b4a6128-dde0-4cea-b0c2-86e6122a638a {{(pid=63088) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 976.516471] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.249s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.517076] env[63088]: DEBUG nova.compute.manager [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 976.519802] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.151s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.521486] env[63088]: INFO nova.compute.claims [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: 13388466-4adb-4d56-9fc1-e3f5dc516077] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 976.524721] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-79c1771e-6b38-47c6-b2c7-527b474cbc1c tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 818704d4-2010-4492-8e51-28ad0bd6f570] Creating Snapshot of the VM instance {{(pid=63088) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 976.525235] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-b5e3c2a4-bbf5-4c52-acff-8debf0e820f4 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.533308] env[63088]: DEBUG oslo_vmware.api [None req-79c1771e-6b38-47c6-b2c7-527b474cbc1c tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Waiting for the task: (returnval){ [ 976.533308] env[63088]: value = "task-1285331" [ 976.533308] env[63088]: _type = "Task" [ 976.533308] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.541564] env[63088]: DEBUG oslo_vmware.api [None req-79c1771e-6b38-47c6-b2c7-527b474cbc1c tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285331, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.712313] env[63088]: DEBUG oslo_concurrency.lockutils [None req-47d36dda-27bb-484d-9238-4b4207ac4046 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Lock "interface-f582da15-dfc3-45e1-a995-9dd0c9533869-051455f3-4c94-4050-8610-bd8ace9f3988" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 9.356s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.731224] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b50fa084-c8dc-450a-9802-45fa9f2a3188 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquiring lock "66178bd0-7ddf-458c-98a2-3519e536dc63" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.731437] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b50fa084-c8dc-450a-9802-45fa9f2a3188 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "66178bd0-7ddf-458c-98a2-3519e536dc63" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.731633] env[63088]: DEBUG nova.compute.manager [None req-b50fa084-c8dc-450a-9802-45fa9f2a3188 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 66178bd0-7ddf-458c-98a2-3519e536dc63] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 976.732580] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db890b13-1f5d-4a8c-a383-ca4cb279f6df {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.740889] env[63088]: DEBUG nova.compute.manager [None req-b50fa084-c8dc-450a-9802-45fa9f2a3188 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 66178bd0-7ddf-458c-98a2-3519e536dc63] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=63088) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 976.741394] env[63088]: DEBUG nova.objects.instance [None req-b50fa084-c8dc-450a-9802-45fa9f2a3188 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lazy-loading 'flavor' on Instance uuid 66178bd0-7ddf-458c-98a2-3519e536dc63 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 977.025990] env[63088]: DEBUG nova.compute.utils [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 977.030010] env[63088]: DEBUG nova.compute.manager [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 977.030010] env[63088]: DEBUG nova.network.neutron [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 977.043837] env[63088]: DEBUG oslo_vmware.api [None req-79c1771e-6b38-47c6-b2c7-527b474cbc1c tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285331, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.105287] env[63088]: DEBUG nova.policy [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '787d7bd9f8e7490c9f24ff0e159bc98f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'dace8b5181b84623b08f903d12dfd31e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 977.247074] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-b50fa084-c8dc-450a-9802-45fa9f2a3188 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 66178bd0-7ddf-458c-98a2-3519e536dc63] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 977.247381] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2220b28c-0f1d-463a-8bac-5043f6c3f4c5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.275884] env[63088]: DEBUG oslo_vmware.api [None req-b50fa084-c8dc-450a-9802-45fa9f2a3188 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 977.275884] env[63088]: value = "task-1285335" [ 977.275884] env[63088]: _type = "Task" [ 977.275884] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.284698] env[63088]: DEBUG oslo_vmware.api [None req-b50fa084-c8dc-450a-9802-45fa9f2a3188 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285335, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.434520] env[63088]: DEBUG nova.network.neutron [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] Successfully created port: 23b3723c-8ff5-467b-b65b-7eb1ee341733 {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 977.532959] env[63088]: DEBUG nova.compute.manager [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 977.548697] env[63088]: DEBUG oslo_vmware.api [None req-79c1771e-6b38-47c6-b2c7-527b474cbc1c tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285331, 'name': CreateSnapshot_Task, 'duration_secs': 0.768052} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.548978] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-79c1771e-6b38-47c6-b2c7-527b474cbc1c tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 818704d4-2010-4492-8e51-28ad0bd6f570] Created Snapshot of the VM instance {{(pid=63088) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 977.550159] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a03abd7-180a-4ce1-b565-5445fffbf04f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.742225] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00e5025c-25a8-4968-9435-12ade6392b6e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.749347] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f439b00e-e136-4b8b-802d-6084a7f692bc {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.788071] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-860764fa-d1e9-4ba3-a756-415b7b65c86b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.801053] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c34947f-bc3c-4a98-9705-bcb70d3ef10e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.804940] env[63088]: DEBUG oslo_vmware.api [None req-b50fa084-c8dc-450a-9802-45fa9f2a3188 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285335, 'name': PowerOffVM_Task, 'duration_secs': 0.382068} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.805370] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-b50fa084-c8dc-450a-9802-45fa9f2a3188 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 66178bd0-7ddf-458c-98a2-3519e536dc63] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 977.805664] env[63088]: DEBUG nova.compute.manager [None req-b50fa084-c8dc-450a-9802-45fa9f2a3188 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 66178bd0-7ddf-458c-98a2-3519e536dc63] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 977.806785] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a6c7646-2d50-4d4c-abf5-a50c685b3adb {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.816765] env[63088]: DEBUG nova.compute.provider_tree [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 978.068266] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-79c1771e-6b38-47c6-b2c7-527b474cbc1c tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 818704d4-2010-4492-8e51-28ad0bd6f570] Creating linked-clone VM from snapshot {{(pid=63088) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 978.068632] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-dd38a5b3-2f64-4180-8ad7-2dfc93fbe182 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.072030] env[63088]: DEBUG oslo_concurrency.lockutils [None req-784c107b-9a9f-49f2-ad33-32d44d1bf7ff tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquiring lock "interface-f582da15-dfc3-45e1-a995-9dd0c9533869-051455f3-4c94-4050-8610-bd8ace9f3988" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 978.072270] env[63088]: DEBUG oslo_concurrency.lockutils [None req-784c107b-9a9f-49f2-ad33-32d44d1bf7ff tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Lock "interface-f582da15-dfc3-45e1-a995-9dd0c9533869-051455f3-4c94-4050-8610-bd8ace9f3988" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 978.080944] env[63088]: DEBUG oslo_vmware.api [None req-79c1771e-6b38-47c6-b2c7-527b474cbc1c tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Waiting for the task: (returnval){ [ 978.080944] env[63088]: value = "task-1285336" [ 978.080944] env[63088]: _type = "Task" [ 978.080944] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.090736] env[63088]: DEBUG oslo_vmware.api [None req-79c1771e-6b38-47c6-b2c7-527b474cbc1c tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285336, 'name': CloneVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.321058] env[63088]: DEBUG nova.scheduler.client.report [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 978.328479] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b50fa084-c8dc-450a-9802-45fa9f2a3188 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "66178bd0-7ddf-458c-98a2-3519e536dc63" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.597s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.546629] env[63088]: DEBUG nova.compute.manager [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 978.572639] env[63088]: DEBUG nova.virt.hardware [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 978.573024] env[63088]: DEBUG nova.virt.hardware [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 978.573285] env[63088]: DEBUG nova.virt.hardware [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 978.573585] env[63088]: DEBUG nova.virt.hardware [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 978.573833] env[63088]: DEBUG nova.virt.hardware [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 978.574095] env[63088]: DEBUG nova.virt.hardware [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 978.574421] env[63088]: DEBUG nova.virt.hardware [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 978.574692] env[63088]: DEBUG nova.virt.hardware [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 978.574955] env[63088]: DEBUG nova.virt.hardware [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 978.575245] env[63088]: DEBUG nova.virt.hardware [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 978.575517] env[63088]: DEBUG nova.virt.hardware [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 978.576200] env[63088]: DEBUG oslo_concurrency.lockutils [None req-784c107b-9a9f-49f2-ad33-32d44d1bf7ff tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquiring lock "f582da15-dfc3-45e1-a995-9dd0c9533869" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 978.576551] env[63088]: DEBUG oslo_concurrency.lockutils [None req-784c107b-9a9f-49f2-ad33-32d44d1bf7ff tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquired lock "f582da15-dfc3-45e1-a995-9dd0c9533869" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 978.577621] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5027d715-cbc1-48bb-9053-4b258ce69c07 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.582365] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-979ed298-42f2-455e-b240-607cde8319c0 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.610726] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de6845de-faba-42c6-908f-4b98288c3190 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.614835] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9382f75a-2542-4c96-b611-119f0e2274e6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.619276] env[63088]: DEBUG oslo_vmware.api [None req-79c1771e-6b38-47c6-b2c7-527b474cbc1c tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285336, 'name': CloneVM_Task} progress is 94%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.645672] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-784c107b-9a9f-49f2-ad33-32d44d1bf7ff tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Reconfiguring VM to detach interface {{(pid=63088) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1974}} [ 978.654894] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-30cd764d-4b96-41ac-a055-4b01b5ecce0e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.678750] env[63088]: DEBUG oslo_vmware.api [None req-784c107b-9a9f-49f2-ad33-32d44d1bf7ff tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Waiting for the task: (returnval){ [ 978.678750] env[63088]: value = "task-1285337" [ 978.678750] env[63088]: _type = "Task" [ 978.678750] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.688168] env[63088]: DEBUG oslo_vmware.api [None req-784c107b-9a9f-49f2-ad33-32d44d1bf7ff tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285337, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.825573] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.306s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.826410] env[63088]: DEBUG nova.compute.manager [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: 13388466-4adb-4d56-9fc1-e3f5dc516077] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 978.829293] env[63088]: DEBUG oslo_concurrency.lockutils [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.009s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 978.830730] env[63088]: INFO nova.compute.claims [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 978.924936] env[63088]: DEBUG nova.compute.manager [req-4ed837a1-057c-4e0a-8ed4-a754feea5dd9 req-8bf4aeb8-d151-43e5-abd9-c9c739fe9763 service nova] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] Received event network-vif-plugged-23b3723c-8ff5-467b-b65b-7eb1ee341733 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 978.925181] env[63088]: DEBUG oslo_concurrency.lockutils [req-4ed837a1-057c-4e0a-8ed4-a754feea5dd9 req-8bf4aeb8-d151-43e5-abd9-c9c739fe9763 service nova] Acquiring lock "69645f50-46fc-4c15-9b39-1721f7636e31-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 978.925399] env[63088]: DEBUG oslo_concurrency.lockutils [req-4ed837a1-057c-4e0a-8ed4-a754feea5dd9 req-8bf4aeb8-d151-43e5-abd9-c9c739fe9763 service nova] Lock "69645f50-46fc-4c15-9b39-1721f7636e31-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 978.925565] env[63088]: DEBUG oslo_concurrency.lockutils [req-4ed837a1-057c-4e0a-8ed4-a754feea5dd9 req-8bf4aeb8-d151-43e5-abd9-c9c739fe9763 service nova] Lock "69645f50-46fc-4c15-9b39-1721f7636e31-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.925729] env[63088]: DEBUG nova.compute.manager [req-4ed837a1-057c-4e0a-8ed4-a754feea5dd9 req-8bf4aeb8-d151-43e5-abd9-c9c739fe9763 service nova] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] No waiting events found dispatching network-vif-plugged-23b3723c-8ff5-467b-b65b-7eb1ee341733 {{(pid=63088) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 978.925901] env[63088]: WARNING nova.compute.manager [req-4ed837a1-057c-4e0a-8ed4-a754feea5dd9 req-8bf4aeb8-d151-43e5-abd9-c9c739fe9763 service nova] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] Received unexpected event network-vif-plugged-23b3723c-8ff5-467b-b65b-7eb1ee341733 for instance with vm_state building and task_state spawning. [ 979.012856] env[63088]: DEBUG nova.network.neutron [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] Successfully updated port: 23b3723c-8ff5-467b-b65b-7eb1ee341733 {{(pid=63088) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 979.095765] env[63088]: DEBUG oslo_vmware.api [None req-79c1771e-6b38-47c6-b2c7-527b474cbc1c tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285336, 'name': CloneVM_Task} progress is 95%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.172306] env[63088]: DEBUG oslo_concurrency.lockutils [None req-56ee9ff4-78f5-42d5-895a-1b5958fb89c8 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquiring lock "66178bd0-7ddf-458c-98a2-3519e536dc63" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 979.172586] env[63088]: DEBUG oslo_concurrency.lockutils [None req-56ee9ff4-78f5-42d5-895a-1b5958fb89c8 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "66178bd0-7ddf-458c-98a2-3519e536dc63" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 979.172805] env[63088]: DEBUG oslo_concurrency.lockutils [None req-56ee9ff4-78f5-42d5-895a-1b5958fb89c8 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquiring lock "66178bd0-7ddf-458c-98a2-3519e536dc63-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 979.172988] env[63088]: DEBUG oslo_concurrency.lockutils [None req-56ee9ff4-78f5-42d5-895a-1b5958fb89c8 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "66178bd0-7ddf-458c-98a2-3519e536dc63-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 979.173217] env[63088]: DEBUG oslo_concurrency.lockutils [None req-56ee9ff4-78f5-42d5-895a-1b5958fb89c8 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "66178bd0-7ddf-458c-98a2-3519e536dc63-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.175358] env[63088]: INFO nova.compute.manager [None req-56ee9ff4-78f5-42d5-895a-1b5958fb89c8 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 66178bd0-7ddf-458c-98a2-3519e536dc63] Terminating instance [ 979.177180] env[63088]: DEBUG nova.compute.manager [None req-56ee9ff4-78f5-42d5-895a-1b5958fb89c8 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 66178bd0-7ddf-458c-98a2-3519e536dc63] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 979.177434] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-56ee9ff4-78f5-42d5-895a-1b5958fb89c8 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 66178bd0-7ddf-458c-98a2-3519e536dc63] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 979.178424] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6e8fbf7-83be-48fa-80ab-23f93b538e43 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.194868] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-56ee9ff4-78f5-42d5-895a-1b5958fb89c8 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 66178bd0-7ddf-458c-98a2-3519e536dc63] Unregistering the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 979.195166] env[63088]: DEBUG oslo_vmware.api [None req-784c107b-9a9f-49f2-ad33-32d44d1bf7ff tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285337, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.195463] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-90d23d73-ac2b-4a7c-8a9e-b9a0ca673480 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.278622] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-56ee9ff4-78f5-42d5-895a-1b5958fb89c8 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 66178bd0-7ddf-458c-98a2-3519e536dc63] Unregistered the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 979.278622] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-56ee9ff4-78f5-42d5-895a-1b5958fb89c8 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 66178bd0-7ddf-458c-98a2-3519e536dc63] Deleting contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 979.278804] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-56ee9ff4-78f5-42d5-895a-1b5958fb89c8 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Deleting the datastore file [datastore1] 66178bd0-7ddf-458c-98a2-3519e536dc63 {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 979.279117] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1e291794-fe95-4837-9d90-538607088988 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.286372] env[63088]: DEBUG oslo_vmware.api [None req-56ee9ff4-78f5-42d5-895a-1b5958fb89c8 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 979.286372] env[63088]: value = "task-1285340" [ 979.286372] env[63088]: _type = "Task" [ 979.286372] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.294957] env[63088]: DEBUG oslo_vmware.api [None req-56ee9ff4-78f5-42d5-895a-1b5958fb89c8 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285340, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.335489] env[63088]: DEBUG nova.compute.utils [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 979.338740] env[63088]: DEBUG nova.compute.manager [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: 13388466-4adb-4d56-9fc1-e3f5dc516077] Not allocating networking since 'none' was specified. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 979.515225] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquiring lock "refresh_cache-69645f50-46fc-4c15-9b39-1721f7636e31" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 979.515401] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquired lock "refresh_cache-69645f50-46fc-4c15-9b39-1721f7636e31" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 979.515518] env[63088]: DEBUG nova.network.neutron [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 979.597896] env[63088]: DEBUG oslo_vmware.api [None req-79c1771e-6b38-47c6-b2c7-527b474cbc1c tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285336, 'name': CloneVM_Task, 'duration_secs': 1.194285} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.598179] env[63088]: INFO nova.virt.vmwareapi.vmops [None req-79c1771e-6b38-47c6-b2c7-527b474cbc1c tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 818704d4-2010-4492-8e51-28ad0bd6f570] Created linked-clone VM from snapshot [ 979.599212] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f07012f4-4525-4e03-820b-0ba9a2b54242 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.607204] env[63088]: DEBUG nova.virt.vmwareapi.images [None req-79c1771e-6b38-47c6-b2c7-527b474cbc1c tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 818704d4-2010-4492-8e51-28ad0bd6f570] Uploading image 635d89fb-eeaf-4b08-a8ea-7be9b1c80e8c {{(pid=63088) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 979.620074] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-79c1771e-6b38-47c6-b2c7-527b474cbc1c tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 818704d4-2010-4492-8e51-28ad0bd6f570] Destroying the VM {{(pid=63088) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 979.620401] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-b8a718e5-92e7-414b-88cf-010b89e086d6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.627715] env[63088]: DEBUG oslo_vmware.api [None req-79c1771e-6b38-47c6-b2c7-527b474cbc1c tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Waiting for the task: (returnval){ [ 979.627715] env[63088]: value = "task-1285341" [ 979.627715] env[63088]: _type = "Task" [ 979.627715] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.636830] env[63088]: DEBUG oslo_vmware.api [None req-79c1771e-6b38-47c6-b2c7-527b474cbc1c tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285341, 'name': Destroy_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.692815] env[63088]: DEBUG oslo_vmware.api [None req-784c107b-9a9f-49f2-ad33-32d44d1bf7ff tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285337, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.796100] env[63088]: DEBUG oslo_vmware.api [None req-56ee9ff4-78f5-42d5-895a-1b5958fb89c8 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285340, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.269296} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.796383] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-56ee9ff4-78f5-42d5-895a-1b5958fb89c8 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 979.796678] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-56ee9ff4-78f5-42d5-895a-1b5958fb89c8 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 66178bd0-7ddf-458c-98a2-3519e536dc63] Deleted contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 979.796759] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-56ee9ff4-78f5-42d5-895a-1b5958fb89c8 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 66178bd0-7ddf-458c-98a2-3519e536dc63] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 979.796929] env[63088]: INFO nova.compute.manager [None req-56ee9ff4-78f5-42d5-895a-1b5958fb89c8 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 66178bd0-7ddf-458c-98a2-3519e536dc63] Took 0.62 seconds to destroy the instance on the hypervisor. [ 979.797252] env[63088]: DEBUG oslo.service.loopingcall [None req-56ee9ff4-78f5-42d5-895a-1b5958fb89c8 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 979.797530] env[63088]: DEBUG nova.compute.manager [-] [instance: 66178bd0-7ddf-458c-98a2-3519e536dc63] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 979.797641] env[63088]: DEBUG nova.network.neutron [-] [instance: 66178bd0-7ddf-458c-98a2-3519e536dc63] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 979.839301] env[63088]: DEBUG nova.compute.manager [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: 13388466-4adb-4d56-9fc1-e3f5dc516077] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 980.016512] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2aaefa64-12aa-4534-bc74-0def4f7a9ca3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.027553] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bf94792-fb6d-4d2b-aa05-c39eaba7aae3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.059224] env[63088]: DEBUG nova.network.neutron [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 980.063100] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3204032d-a621-4f3f-a39d-c5866e486cf9 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.072278] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9d16a42-6873-42fe-ba33-b20ac2d1e3be {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.088576] env[63088]: DEBUG nova.compute.provider_tree [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 980.139354] env[63088]: DEBUG oslo_vmware.api [None req-79c1771e-6b38-47c6-b2c7-527b474cbc1c tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285341, 'name': Destroy_Task, 'duration_secs': 0.36841} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.139731] env[63088]: INFO nova.virt.vmwareapi.vm_util [None req-79c1771e-6b38-47c6-b2c7-527b474cbc1c tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 818704d4-2010-4492-8e51-28ad0bd6f570] Destroyed the VM [ 980.139911] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-79c1771e-6b38-47c6-b2c7-527b474cbc1c tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 818704d4-2010-4492-8e51-28ad0bd6f570] Deleting Snapshot of the VM instance {{(pid=63088) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 980.140198] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-e4b9f421-92e7-4f63-9e9d-f75b16b68bcb {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.150704] env[63088]: DEBUG oslo_vmware.api [None req-79c1771e-6b38-47c6-b2c7-527b474cbc1c tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Waiting for the task: (returnval){ [ 980.150704] env[63088]: value = "task-1285342" [ 980.150704] env[63088]: _type = "Task" [ 980.150704] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.159946] env[63088]: DEBUG oslo_vmware.api [None req-79c1771e-6b38-47c6-b2c7-527b474cbc1c tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285342, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.194364] env[63088]: DEBUG oslo_vmware.api [None req-784c107b-9a9f-49f2-ad33-32d44d1bf7ff tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285337, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.484902] env[63088]: DEBUG nova.network.neutron [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] Updating instance_info_cache with network_info: [{"id": "23b3723c-8ff5-467b-b65b-7eb1ee341733", "address": "fa:16:3e:94:55:ca", "network": {"id": "dfcbe153-bc01-4362-9247-de9f0b1c847b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-841757397-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dace8b5181b84623b08f903d12dfd31e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c68b7663-4f0e-47f0-ac7f-40c6d952f7bb", "external-id": "nsx-vlan-transportzone-696", "segmentation_id": 696, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap23b3723c-8f", "ovs_interfaceid": "23b3723c-8ff5-467b-b65b-7eb1ee341733", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 980.566743] env[63088]: DEBUG nova.network.neutron [-] [instance: 66178bd0-7ddf-458c-98a2-3519e536dc63] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 980.593146] env[63088]: DEBUG nova.scheduler.client.report [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 980.663339] env[63088]: DEBUG oslo_vmware.api [None req-79c1771e-6b38-47c6-b2c7-527b474cbc1c tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285342, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.694497] env[63088]: DEBUG oslo_vmware.api [None req-784c107b-9a9f-49f2-ad33-32d44d1bf7ff tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285337, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.858061] env[63088]: DEBUG nova.compute.manager [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: 13388466-4adb-4d56-9fc1-e3f5dc516077] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 980.886994] env[63088]: DEBUG nova.virt.hardware [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 980.887281] env[63088]: DEBUG nova.virt.hardware [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 980.887466] env[63088]: DEBUG nova.virt.hardware [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 980.887651] env[63088]: DEBUG nova.virt.hardware [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 980.887805] env[63088]: DEBUG nova.virt.hardware [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 980.887952] env[63088]: DEBUG nova.virt.hardware [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 980.888185] env[63088]: DEBUG nova.virt.hardware [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 980.888387] env[63088]: DEBUG nova.virt.hardware [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 980.888709] env[63088]: DEBUG nova.virt.hardware [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 980.888917] env[63088]: DEBUG nova.virt.hardware [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 980.889115] env[63088]: DEBUG nova.virt.hardware [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 980.890166] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48fe6966-1c71-48e4-bbaf-72f2ea8da695 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.900114] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfacb72f-4226-48ca-9bff-d1043558800e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.915574] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: 13388466-4adb-4d56-9fc1-e3f5dc516077] Instance VIF info [] {{(pid=63088) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 980.921323] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Creating folder: Project (63990c6c10a4468b890592c099bbd88c). Parent ref: group-v275816. {{(pid=63088) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 980.921699] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-81d18590-2080-4867-9923-e8b4d1f2799e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.936754] env[63088]: INFO nova.virt.vmwareapi.vm_util [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Created folder: Project (63990c6c10a4468b890592c099bbd88c) in parent group-v275816. [ 980.937155] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Creating folder: Instances. Parent ref: group-v275935. {{(pid=63088) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 980.937155] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ce103960-ddcc-4dae-83c4-3233435d0de4 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.950058] env[63088]: INFO nova.virt.vmwareapi.vm_util [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Created folder: Instances in parent group-v275935. [ 980.951029] env[63088]: DEBUG oslo.service.loopingcall [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 980.951455] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 13388466-4adb-4d56-9fc1-e3f5dc516077] Creating VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 980.951817] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-697862f7-1539-4fbf-945f-f5d656553c2b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.981186] env[63088]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 980.981186] env[63088]: value = "task-1285345" [ 980.981186] env[63088]: _type = "Task" [ 980.981186] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.987625] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Releasing lock "refresh_cache-69645f50-46fc-4c15-9b39-1721f7636e31" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 980.987957] env[63088]: DEBUG nova.compute.manager [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] Instance network_info: |[{"id": "23b3723c-8ff5-467b-b65b-7eb1ee341733", "address": "fa:16:3e:94:55:ca", "network": {"id": "dfcbe153-bc01-4362-9247-de9f0b1c847b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-841757397-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dace8b5181b84623b08f903d12dfd31e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c68b7663-4f0e-47f0-ac7f-40c6d952f7bb", "external-id": "nsx-vlan-transportzone-696", "segmentation_id": 696, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap23b3723c-8f", "ovs_interfaceid": "23b3723c-8ff5-467b-b65b-7eb1ee341733", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 980.988435] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:94:55:ca', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c68b7663-4f0e-47f0-ac7f-40c6d952f7bb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '23b3723c-8ff5-467b-b65b-7eb1ee341733', 'vif_model': 'vmxnet3'}] {{(pid=63088) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 980.996616] env[63088]: DEBUG oslo.service.loopingcall [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 981.000843] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] Creating VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 981.001103] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285345, 'name': CreateVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.001348] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a460f843-3bee-457f-97f4-740e4e66124e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.028094] env[63088]: DEBUG nova.compute.manager [req-c7cc1a0f-84b8-4e80-a136-575b862ec2dc req-579c331b-7caf-4998-8578-6137afd3be06 service nova] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] Received event network-changed-23b3723c-8ff5-467b-b65b-7eb1ee341733 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 981.028313] env[63088]: DEBUG nova.compute.manager [req-c7cc1a0f-84b8-4e80-a136-575b862ec2dc req-579c331b-7caf-4998-8578-6137afd3be06 service nova] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] Refreshing instance network info cache due to event network-changed-23b3723c-8ff5-467b-b65b-7eb1ee341733. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 981.028537] env[63088]: DEBUG oslo_concurrency.lockutils [req-c7cc1a0f-84b8-4e80-a136-575b862ec2dc req-579c331b-7caf-4998-8578-6137afd3be06 service nova] Acquiring lock "refresh_cache-69645f50-46fc-4c15-9b39-1721f7636e31" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 981.028684] env[63088]: DEBUG oslo_concurrency.lockutils [req-c7cc1a0f-84b8-4e80-a136-575b862ec2dc req-579c331b-7caf-4998-8578-6137afd3be06 service nova] Acquired lock "refresh_cache-69645f50-46fc-4c15-9b39-1721f7636e31" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 981.028848] env[63088]: DEBUG nova.network.neutron [req-c7cc1a0f-84b8-4e80-a136-575b862ec2dc req-579c331b-7caf-4998-8578-6137afd3be06 service nova] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] Refreshing network info cache for port 23b3723c-8ff5-467b-b65b-7eb1ee341733 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 981.030200] env[63088]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 981.030200] env[63088]: value = "task-1285346" [ 981.030200] env[63088]: _type = "Task" [ 981.030200] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.041444] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285346, 'name': CreateVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.070120] env[63088]: INFO nova.compute.manager [-] [instance: 66178bd0-7ddf-458c-98a2-3519e536dc63] Took 1.27 seconds to deallocate network for instance. [ 981.102097] env[63088]: DEBUG oslo_concurrency.lockutils [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.272s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.102613] env[63088]: DEBUG nova.compute.manager [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 981.105791] env[63088]: DEBUG oslo_concurrency.lockutils [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.883s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.107337] env[63088]: INFO nova.compute.claims [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: c4fb31b2-1301-4a6e-bef2-8ea462caee0d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 981.162699] env[63088]: DEBUG oslo_vmware.api [None req-79c1771e-6b38-47c6-b2c7-527b474cbc1c tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285342, 'name': RemoveSnapshot_Task, 'duration_secs': 0.737041} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.163209] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-79c1771e-6b38-47c6-b2c7-527b474cbc1c tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 818704d4-2010-4492-8e51-28ad0bd6f570] Deleted Snapshot of the VM instance {{(pid=63088) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 981.197199] env[63088]: DEBUG oslo_vmware.api [None req-784c107b-9a9f-49f2-ad33-32d44d1bf7ff tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285337, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.491607] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285345, 'name': CreateVM_Task, 'duration_secs': 0.314267} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.491821] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 13388466-4adb-4d56-9fc1-e3f5dc516077] Created VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 981.492241] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 981.492404] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 981.492749] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 981.493019] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2852c3e5-3044-4800-9950-1456c9e63e84 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.498441] env[63088]: DEBUG oslo_vmware.api [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Waiting for the task: (returnval){ [ 981.498441] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]5231b805-171e-031a-8ca8-164e4dff59e3" [ 981.498441] env[63088]: _type = "Task" [ 981.498441] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.507878] env[63088]: DEBUG oslo_vmware.api [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]5231b805-171e-031a-8ca8-164e4dff59e3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.543501] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285346, 'name': CreateVM_Task, 'duration_secs': 0.427647} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.543686] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] Created VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 981.544373] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 981.576592] env[63088]: DEBUG oslo_concurrency.lockutils [None req-56ee9ff4-78f5-42d5-895a-1b5958fb89c8 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 981.611774] env[63088]: DEBUG nova.compute.utils [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 981.615012] env[63088]: DEBUG nova.compute.manager [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3] Not allocating networking since 'none' was specified. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 981.670376] env[63088]: WARNING nova.compute.manager [None req-79c1771e-6b38-47c6-b2c7-527b474cbc1c tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 818704d4-2010-4492-8e51-28ad0bd6f570] Image not found during snapshot: nova.exception.ImageNotFound: Image 635d89fb-eeaf-4b08-a8ea-7be9b1c80e8c could not be found. [ 981.697591] env[63088]: DEBUG oslo_vmware.api [None req-784c107b-9a9f-49f2-ad33-32d44d1bf7ff tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285337, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.814367] env[63088]: DEBUG nova.network.neutron [req-c7cc1a0f-84b8-4e80-a136-575b862ec2dc req-579c331b-7caf-4998-8578-6137afd3be06 service nova] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] Updated VIF entry in instance network info cache for port 23b3723c-8ff5-467b-b65b-7eb1ee341733. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 981.814839] env[63088]: DEBUG nova.network.neutron [req-c7cc1a0f-84b8-4e80-a136-575b862ec2dc req-579c331b-7caf-4998-8578-6137afd3be06 service nova] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] Updating instance_info_cache with network_info: [{"id": "23b3723c-8ff5-467b-b65b-7eb1ee341733", "address": "fa:16:3e:94:55:ca", "network": {"id": "dfcbe153-bc01-4362-9247-de9f0b1c847b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-841757397-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dace8b5181b84623b08f903d12dfd31e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c68b7663-4f0e-47f0-ac7f-40c6d952f7bb", "external-id": "nsx-vlan-transportzone-696", "segmentation_id": 696, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap23b3723c-8f", "ovs_interfaceid": "23b3723c-8ff5-467b-b65b-7eb1ee341733", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 982.009904] env[63088]: DEBUG oslo_vmware.api [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]5231b805-171e-031a-8ca8-164e4dff59e3, 'name': SearchDatastore_Task, 'duration_secs': 0.034328} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.009904] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 982.010180] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: 13388466-4adb-4d56-9fc1-e3f5dc516077] Processing image 1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 982.010463] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 982.010616] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 982.010804] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 982.011112] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 982.011465] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 982.011700] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8aab2555-90a8-4182-8589-4ecdd492f832 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.013570] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-65954937-f394-454f-8dfb-28a11c9a7627 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.019513] env[63088]: DEBUG oslo_vmware.api [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for the task: (returnval){ [ 982.019513] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52dcc963-d143-0dde-d501-6758d0d1d788" [ 982.019513] env[63088]: _type = "Task" [ 982.019513] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.025854] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 982.025854] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 982.026963] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0fa8616b-2d85-4e9a-ab87-b12c8ed05e05 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.032679] env[63088]: DEBUG oslo_vmware.api [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52dcc963-d143-0dde-d501-6758d0d1d788, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.036189] env[63088]: DEBUG oslo_vmware.api [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Waiting for the task: (returnval){ [ 982.036189] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52ddc1c2-649e-75db-28df-8aebf00a7c3b" [ 982.036189] env[63088]: _type = "Task" [ 982.036189] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.042044] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 982.045606] env[63088]: DEBUG oslo_vmware.api [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52ddc1c2-649e-75db-28df-8aebf00a7c3b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.116054] env[63088]: DEBUG nova.compute.manager [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 982.201193] env[63088]: DEBUG oslo_vmware.api [None req-784c107b-9a9f-49f2-ad33-32d44d1bf7ff tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285337, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.269530] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0817a56-1145-4a16-8c64-8180e7efde00 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.277756] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0527d70-1f2a-4fb4-b152-4038c8ee44c8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.307554] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2aebf90f-192a-4ca6-a440-81ac14f1f782 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.315520] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad4bc7b2-10da-4eda-8acb-613e387ad745 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.321268] env[63088]: DEBUG oslo_concurrency.lockutils [req-c7cc1a0f-84b8-4e80-a136-575b862ec2dc req-579c331b-7caf-4998-8578-6137afd3be06 service nova] Releasing lock "refresh_cache-69645f50-46fc-4c15-9b39-1721f7636e31" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 982.321538] env[63088]: DEBUG nova.compute.manager [req-c7cc1a0f-84b8-4e80-a136-575b862ec2dc req-579c331b-7caf-4998-8578-6137afd3be06 service nova] [instance: 66178bd0-7ddf-458c-98a2-3519e536dc63] Received event network-vif-deleted-725d6827-6db3-4f39-8374-df9d333b77bb {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 982.332065] env[63088]: DEBUG nova.compute.provider_tree [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 982.531370] env[63088]: DEBUG oslo_vmware.api [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52dcc963-d143-0dde-d501-6758d0d1d788, 'name': SearchDatastore_Task, 'duration_secs': 0.019926} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.531752] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 982.531955] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] Processing image 1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 982.532182] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 982.546624] env[63088]: DEBUG oslo_vmware.api [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52ddc1c2-649e-75db-28df-8aebf00a7c3b, 'name': SearchDatastore_Task, 'duration_secs': 0.009466} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.547398] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-131a7202-d008-48dc-80ba-c43f0709f884 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.552859] env[63088]: DEBUG oslo_vmware.api [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Waiting for the task: (returnval){ [ 982.552859] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52a5529b-8532-ed8d-ccad-0fe7ba73ad77" [ 982.552859] env[63088]: _type = "Task" [ 982.552859] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.560676] env[63088]: DEBUG oslo_vmware.api [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52a5529b-8532-ed8d-ccad-0fe7ba73ad77, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.689481] env[63088]: DEBUG oslo_concurrency.lockutils [None req-82ce7dd7-1f78-4f20-808f-f5b9d77ac35d tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Acquiring lock "818704d4-2010-4492-8e51-28ad0bd6f570" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 982.689745] env[63088]: DEBUG oslo_concurrency.lockutils [None req-82ce7dd7-1f78-4f20-808f-f5b9d77ac35d tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Lock "818704d4-2010-4492-8e51-28ad0bd6f570" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 982.689950] env[63088]: DEBUG oslo_concurrency.lockutils [None req-82ce7dd7-1f78-4f20-808f-f5b9d77ac35d tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Acquiring lock "818704d4-2010-4492-8e51-28ad0bd6f570-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 982.690156] env[63088]: DEBUG oslo_concurrency.lockutils [None req-82ce7dd7-1f78-4f20-808f-f5b9d77ac35d tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Lock "818704d4-2010-4492-8e51-28ad0bd6f570-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 982.690357] env[63088]: DEBUG oslo_concurrency.lockutils [None req-82ce7dd7-1f78-4f20-808f-f5b9d77ac35d tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Lock "818704d4-2010-4492-8e51-28ad0bd6f570-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 982.695597] env[63088]: INFO nova.compute.manager [None req-82ce7dd7-1f78-4f20-808f-f5b9d77ac35d tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 818704d4-2010-4492-8e51-28ad0bd6f570] Terminating instance [ 982.697718] env[63088]: DEBUG nova.compute.manager [None req-82ce7dd7-1f78-4f20-808f-f5b9d77ac35d tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 818704d4-2010-4492-8e51-28ad0bd6f570] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 982.697929] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-82ce7dd7-1f78-4f20-808f-f5b9d77ac35d tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 818704d4-2010-4492-8e51-28ad0bd6f570] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 982.698691] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b91a8a58-e275-4b15-a21b-b145264bdec4 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.704631] env[63088]: DEBUG oslo_vmware.api [None req-784c107b-9a9f-49f2-ad33-32d44d1bf7ff tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285337, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.709680] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-82ce7dd7-1f78-4f20-808f-f5b9d77ac35d tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 818704d4-2010-4492-8e51-28ad0bd6f570] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 982.709949] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-36716251-3c91-4c38-9ec5-30aa874f2217 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.716262] env[63088]: DEBUG oslo_vmware.api [None req-82ce7dd7-1f78-4f20-808f-f5b9d77ac35d tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Waiting for the task: (returnval){ [ 982.716262] env[63088]: value = "task-1285347" [ 982.716262] env[63088]: _type = "Task" [ 982.716262] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.725320] env[63088]: DEBUG oslo_vmware.api [None req-82ce7dd7-1f78-4f20-808f-f5b9d77ac35d tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285347, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.835321] env[63088]: DEBUG nova.scheduler.client.report [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 983.064130] env[63088]: DEBUG oslo_vmware.api [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52a5529b-8532-ed8d-ccad-0fe7ba73ad77, 'name': SearchDatastore_Task, 'duration_secs': 0.010081} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.064449] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 983.064769] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] 13388466-4adb-4d56-9fc1-e3f5dc516077/13388466-4adb-4d56-9fc1-e3f5dc516077.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 983.065142] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 983.065407] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 983.065699] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d61e4df8-c208-4724-b71c-6668ccde4646 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.067811] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-564b9d6c-5fd6-4bff-aca3-111bc93d13b9 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.075475] env[63088]: DEBUG oslo_vmware.api [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Waiting for the task: (returnval){ [ 983.075475] env[63088]: value = "task-1285348" [ 983.075475] env[63088]: _type = "Task" [ 983.075475] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.079707] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 983.079898] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 983.081109] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8efdf604-3a9c-4b6a-a58f-23d4fed47431 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.086705] env[63088]: DEBUG oslo_vmware.api [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Task: {'id': task-1285348, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.090138] env[63088]: DEBUG oslo_vmware.api [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for the task: (returnval){ [ 983.090138] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]524931be-39bc-9c3d-a910-6f1489924ecf" [ 983.090138] env[63088]: _type = "Task" [ 983.090138] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.098280] env[63088]: DEBUG oslo_vmware.api [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]524931be-39bc-9c3d-a910-6f1489924ecf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.128179] env[63088]: DEBUG nova.compute.manager [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 983.151466] env[63088]: DEBUG nova.virt.hardware [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 983.151734] env[63088]: DEBUG nova.virt.hardware [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 983.151910] env[63088]: DEBUG nova.virt.hardware [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 983.152124] env[63088]: DEBUG nova.virt.hardware [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 983.152289] env[63088]: DEBUG nova.virt.hardware [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 983.152455] env[63088]: DEBUG nova.virt.hardware [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 983.152692] env[63088]: DEBUG nova.virt.hardware [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 983.152858] env[63088]: DEBUG nova.virt.hardware [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 983.153044] env[63088]: DEBUG nova.virt.hardware [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 983.153236] env[63088]: DEBUG nova.virt.hardware [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 983.153440] env[63088]: DEBUG nova.virt.hardware [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 983.154348] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9ad0aaa-6a7b-4d58-aac6-dc9e3ea7e311 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.162934] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41a6656b-1547-4274-896d-d24033bf4618 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.178055] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3] Instance VIF info [] {{(pid=63088) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 983.183490] env[63088]: DEBUG oslo.service.loopingcall [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 983.183784] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3] Creating VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 983.184041] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ba5903e3-7a43-4236-88d7-0aef88a1b969 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.204179] env[63088]: DEBUG oslo_vmware.api [None req-784c107b-9a9f-49f2-ad33-32d44d1bf7ff tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285337, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.205371] env[63088]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 983.205371] env[63088]: value = "task-1285349" [ 983.205371] env[63088]: _type = "Task" [ 983.205371] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.212550] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285349, 'name': CreateVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.224502] env[63088]: DEBUG oslo_vmware.api [None req-82ce7dd7-1f78-4f20-808f-f5b9d77ac35d tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285347, 'name': PowerOffVM_Task, 'duration_secs': 0.189972} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.224761] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-82ce7dd7-1f78-4f20-808f-f5b9d77ac35d tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 818704d4-2010-4492-8e51-28ad0bd6f570] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 983.224944] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-82ce7dd7-1f78-4f20-808f-f5b9d77ac35d tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 818704d4-2010-4492-8e51-28ad0bd6f570] Unregistering the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 983.225251] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9e4ec9c0-d264-4d0e-bf0d-cd1b7a8760c1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.313745] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-82ce7dd7-1f78-4f20-808f-f5b9d77ac35d tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 818704d4-2010-4492-8e51-28ad0bd6f570] Unregistered the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 983.314097] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-82ce7dd7-1f78-4f20-808f-f5b9d77ac35d tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 818704d4-2010-4492-8e51-28ad0bd6f570] Deleting contents of the VM from datastore datastore2 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 983.314395] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-82ce7dd7-1f78-4f20-808f-f5b9d77ac35d tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Deleting the datastore file [datastore2] 818704d4-2010-4492-8e51-28ad0bd6f570 {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 983.314822] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-eddbef9f-9c70-4334-81d0-790588a08034 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.325473] env[63088]: DEBUG oslo_vmware.api [None req-82ce7dd7-1f78-4f20-808f-f5b9d77ac35d tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Waiting for the task: (returnval){ [ 983.325473] env[63088]: value = "task-1285351" [ 983.325473] env[63088]: _type = "Task" [ 983.325473] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.335418] env[63088]: DEBUG oslo_vmware.api [None req-82ce7dd7-1f78-4f20-808f-f5b9d77ac35d tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285351, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.340592] env[63088]: DEBUG oslo_concurrency.lockutils [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.235s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 983.341119] env[63088]: DEBUG nova.compute.manager [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: c4fb31b2-1301-4a6e-bef2-8ea462caee0d] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 983.344026] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bc161c1d-84e4-48a5-a6e5-b5d8ec5b62aa tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 9.927s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 983.586349] env[63088]: DEBUG oslo_vmware.api [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Task: {'id': task-1285348, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.475652} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.586679] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] 13388466-4adb-4d56-9fc1-e3f5dc516077/13388466-4adb-4d56-9fc1-e3f5dc516077.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 983.586903] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: 13388466-4adb-4d56-9fc1-e3f5dc516077] Extending root virtual disk to 1048576 {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 983.587179] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7490ce3b-afb4-4b6e-b4d9-7da7e0dd244d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.595666] env[63088]: DEBUG oslo_vmware.api [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Waiting for the task: (returnval){ [ 983.595666] env[63088]: value = "task-1285352" [ 983.595666] env[63088]: _type = "Task" [ 983.595666] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.602577] env[63088]: DEBUG oslo_vmware.api [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]524931be-39bc-9c3d-a910-6f1489924ecf, 'name': SearchDatastore_Task, 'duration_secs': 0.008894} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.603727] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-78ef55d2-2a38-45a1-8f71-68f46ddfaffa {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.610297] env[63088]: DEBUG oslo_vmware.api [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Task: {'id': task-1285352, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.611747] env[63088]: DEBUG oslo_vmware.api [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for the task: (returnval){ [ 983.611747] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52131e6e-f654-fcee-402f-b98817876b77" [ 983.611747] env[63088]: _type = "Task" [ 983.611747] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.620428] env[63088]: DEBUG oslo_vmware.api [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52131e6e-f654-fcee-402f-b98817876b77, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.706592] env[63088]: DEBUG oslo_vmware.api [None req-784c107b-9a9f-49f2-ad33-32d44d1bf7ff tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285337, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.714861] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285349, 'name': CreateVM_Task} progress is 99%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.835886] env[63088]: DEBUG oslo_vmware.api [None req-82ce7dd7-1f78-4f20-808f-f5b9d77ac35d tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Task: {'id': task-1285351, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.286434} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.836190] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-82ce7dd7-1f78-4f20-808f-f5b9d77ac35d tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 983.836413] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-82ce7dd7-1f78-4f20-808f-f5b9d77ac35d tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 818704d4-2010-4492-8e51-28ad0bd6f570] Deleted contents of the VM from datastore datastore2 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 983.836567] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-82ce7dd7-1f78-4f20-808f-f5b9d77ac35d tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 818704d4-2010-4492-8e51-28ad0bd6f570] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 983.836801] env[63088]: INFO nova.compute.manager [None req-82ce7dd7-1f78-4f20-808f-f5b9d77ac35d tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] [instance: 818704d4-2010-4492-8e51-28ad0bd6f570] Took 1.14 seconds to destroy the instance on the hypervisor. [ 983.836979] env[63088]: DEBUG oslo.service.loopingcall [None req-82ce7dd7-1f78-4f20-808f-f5b9d77ac35d tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 983.837190] env[63088]: DEBUG nova.compute.manager [-] [instance: 818704d4-2010-4492-8e51-28ad0bd6f570] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 983.837287] env[63088]: DEBUG nova.network.neutron [-] [instance: 818704d4-2010-4492-8e51-28ad0bd6f570] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 983.847206] env[63088]: DEBUG nova.compute.utils [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 983.851819] env[63088]: DEBUG nova.compute.manager [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: c4fb31b2-1301-4a6e-bef2-8ea462caee0d] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 983.852151] env[63088]: DEBUG nova.network.neutron [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: c4fb31b2-1301-4a6e-bef2-8ea462caee0d] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 983.958360] env[63088]: DEBUG nova.policy [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '32d5ff9b3ea345e4ac56010aa917e51b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '03e20d606e654362acbe2b36fe499ae3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 984.034279] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b909f0e-bb0d-4eb4-a9b3-744db5392cc9 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.044495] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f18d923-0e90-4e0b-8a89-09cdb3e80a9e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.074940] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfcaf7ba-a514-4572-8f09-1984eca4d6ee {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.084396] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad1c9f6a-cf52-405d-8f6e-c97bad84690c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.100438] env[63088]: DEBUG nova.compute.provider_tree [None req-bc161c1d-84e4-48a5-a6e5-b5d8ec5b62aa tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 984.110095] env[63088]: DEBUG oslo_vmware.api [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Task: {'id': task-1285352, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065163} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.110408] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: 13388466-4adb-4d56-9fc1-e3f5dc516077] Extended root virtual disk {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 984.111267] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d42000b-833c-4510-bbe1-16e00881275d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.116311] env[63088]: DEBUG nova.compute.manager [req-e5a66432-6786-4a34-9091-871ee01a6251 req-543a2649-9286-448f-8b97-12c53a18e146 service nova] [instance: 818704d4-2010-4492-8e51-28ad0bd6f570] Received event network-vif-deleted-2a414d49-0028-4cbf-bc84-04946140a76c {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 984.116441] env[63088]: INFO nova.compute.manager [req-e5a66432-6786-4a34-9091-871ee01a6251 req-543a2649-9286-448f-8b97-12c53a18e146 service nova] [instance: 818704d4-2010-4492-8e51-28ad0bd6f570] Neutron deleted interface 2a414d49-0028-4cbf-bc84-04946140a76c; detaching it from the instance and deleting it from the info cache [ 984.116615] env[63088]: DEBUG nova.network.neutron [req-e5a66432-6786-4a34-9091-871ee01a6251 req-543a2649-9286-448f-8b97-12c53a18e146 service nova] [instance: 818704d4-2010-4492-8e51-28ad0bd6f570] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 984.126983] env[63088]: DEBUG oslo_vmware.api [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52131e6e-f654-fcee-402f-b98817876b77, 'name': SearchDatastore_Task, 'duration_secs': 0.012718} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.134199] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 984.134494] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] 69645f50-46fc-4c15-9b39-1721f7636e31/69645f50-46fc-4c15-9b39-1721f7636e31.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 984.143990] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: 13388466-4adb-4d56-9fc1-e3f5dc516077] Reconfiguring VM instance instance-0000005e to attach disk [datastore1] 13388466-4adb-4d56-9fc1-e3f5dc516077/13388466-4adb-4d56-9fc1-e3f5dc516077.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 984.144290] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d3941a15-60bd-43d0-a117-063edd0680e3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.146924] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5b1b6b65-0d62-4417-bfd0-c11ed880dea1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.167097] env[63088]: DEBUG oslo_vmware.api [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Waiting for the task: (returnval){ [ 984.167097] env[63088]: value = "task-1285354" [ 984.167097] env[63088]: _type = "Task" [ 984.167097] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.168444] env[63088]: DEBUG oslo_vmware.api [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for the task: (returnval){ [ 984.168444] env[63088]: value = "task-1285353" [ 984.168444] env[63088]: _type = "Task" [ 984.168444] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.181570] env[63088]: DEBUG oslo_vmware.api [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285353, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.184251] env[63088]: DEBUG oslo_vmware.api [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Task: {'id': task-1285354, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.207538] env[63088]: DEBUG oslo_vmware.api [None req-784c107b-9a9f-49f2-ad33-32d44d1bf7ff tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285337, 'name': ReconfigVM_Task} progress is 18%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.216450] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285349, 'name': CreateVM_Task, 'duration_secs': 0.518449} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.216640] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3] Created VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 984.217095] env[63088]: DEBUG oslo_concurrency.lockutils [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 984.217259] env[63088]: DEBUG oslo_concurrency.lockutils [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 984.217571] env[63088]: DEBUG oslo_concurrency.lockutils [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 984.217845] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ff892298-bd9e-4dc8-be40-9625fe8c02d0 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.223022] env[63088]: DEBUG oslo_vmware.api [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Waiting for the task: (returnval){ [ 984.223022] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52c0badd-e3f3-4332-091d-9b8b42a7eea7" [ 984.223022] env[63088]: _type = "Task" [ 984.223022] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.232251] env[63088]: DEBUG oslo_vmware.api [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52c0badd-e3f3-4332-091d-9b8b42a7eea7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.263550] env[63088]: DEBUG nova.network.neutron [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: c4fb31b2-1301-4a6e-bef2-8ea462caee0d] Successfully created port: 88bc1e34-07ac-4819-9cd8-d933c1540bb2 {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 984.353290] env[63088]: DEBUG nova.compute.manager [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: c4fb31b2-1301-4a6e-bef2-8ea462caee0d] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 984.582619] env[63088]: DEBUG nova.network.neutron [-] [instance: 818704d4-2010-4492-8e51-28ad0bd6f570] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 984.607033] env[63088]: DEBUG nova.scheduler.client.report [None req-bc161c1d-84e4-48a5-a6e5-b5d8ec5b62aa tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 984.622866] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ba366186-07cd-457d-a016-e153ecad3573 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.634552] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43f33f08-45a8-4c84-bc77-5719b0cf0784 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.668216] env[63088]: DEBUG nova.compute.manager [req-e5a66432-6786-4a34-9091-871ee01a6251 req-543a2649-9286-448f-8b97-12c53a18e146 service nova] [instance: 818704d4-2010-4492-8e51-28ad0bd6f570] Detach interface failed, port_id=2a414d49-0028-4cbf-bc84-04946140a76c, reason: Instance 818704d4-2010-4492-8e51-28ad0bd6f570 could not be found. {{(pid=63088) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 984.680073] env[63088]: DEBUG oslo_vmware.api [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Task: {'id': task-1285354, 'name': ReconfigVM_Task, 'duration_secs': 0.353582} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.680799] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: 13388466-4adb-4d56-9fc1-e3f5dc516077] Reconfigured VM instance instance-0000005e to attach disk [datastore1] 13388466-4adb-4d56-9fc1-e3f5dc516077/13388466-4adb-4d56-9fc1-e3f5dc516077.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 984.681473] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f8e107a6-3aa6-423a-95c0-49a6bfaac580 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.685637] env[63088]: DEBUG oslo_vmware.api [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285353, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.691818] env[63088]: DEBUG oslo_vmware.api [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Waiting for the task: (returnval){ [ 984.691818] env[63088]: value = "task-1285355" [ 984.691818] env[63088]: _type = "Task" [ 984.691818] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.699259] env[63088]: DEBUG oslo_vmware.api [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Task: {'id': task-1285355, 'name': Rename_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.707270] env[63088]: DEBUG oslo_vmware.api [None req-784c107b-9a9f-49f2-ad33-32d44d1bf7ff tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285337, 'name': ReconfigVM_Task, 'duration_secs': 5.781419} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.707508] env[63088]: DEBUG oslo_concurrency.lockutils [None req-784c107b-9a9f-49f2-ad33-32d44d1bf7ff tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Releasing lock "f582da15-dfc3-45e1-a995-9dd0c9533869" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 984.707718] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-784c107b-9a9f-49f2-ad33-32d44d1bf7ff tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Reconfigured VM to detach interface {{(pid=63088) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1984}} [ 984.734318] env[63088]: DEBUG oslo_vmware.api [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52c0badd-e3f3-4332-091d-9b8b42a7eea7, 'name': SearchDatastore_Task, 'duration_secs': 0.023707} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.734689] env[63088]: DEBUG oslo_concurrency.lockutils [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 984.734888] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3] Processing image 1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 984.735163] env[63088]: DEBUG oslo_concurrency.lockutils [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 984.735332] env[63088]: DEBUG oslo_concurrency.lockutils [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 984.735528] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 984.737779] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-94bdf02d-d14b-42af-a97f-f5c4b2f664b2 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.746775] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 984.746981] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 984.747766] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-04dd121a-1de4-4a51-aa6b-f5ead8429031 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.753864] env[63088]: DEBUG oslo_vmware.api [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Waiting for the task: (returnval){ [ 984.753864] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]5228baed-4728-d389-ec58-7a426ec3db90" [ 984.753864] env[63088]: _type = "Task" [ 984.753864] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.764558] env[63088]: DEBUG oslo_vmware.api [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]5228baed-4728-d389-ec58-7a426ec3db90, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.084531] env[63088]: INFO nova.compute.manager [-] [instance: 818704d4-2010-4492-8e51-28ad0bd6f570] Took 1.25 seconds to deallocate network for instance. [ 985.184454] env[63088]: DEBUG oslo_vmware.api [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285353, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.53258} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.184749] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] 69645f50-46fc-4c15-9b39-1721f7636e31/69645f50-46fc-4c15-9b39-1721f7636e31.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 985.185053] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] Extending root virtual disk to 1048576 {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 985.185358] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7db34638-f6e8-44b0-843a-ec6232e3d558 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.193232] env[63088]: DEBUG oslo_vmware.api [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for the task: (returnval){ [ 985.193232] env[63088]: value = "task-1285356" [ 985.193232] env[63088]: _type = "Task" [ 985.193232] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.208900] env[63088]: DEBUG oslo_vmware.api [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285356, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.208900] env[63088]: DEBUG oslo_vmware.api [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Task: {'id': task-1285355, 'name': Rename_Task, 'duration_secs': 0.187708} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.209328] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: 13388466-4adb-4d56-9fc1-e3f5dc516077] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 985.209369] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-66be6a7f-d7a0-4ae5-ba1f-238706f33fa6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.216579] env[63088]: DEBUG oslo_vmware.api [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Waiting for the task: (returnval){ [ 985.216579] env[63088]: value = "task-1285357" [ 985.216579] env[63088]: _type = "Task" [ 985.216579] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.225666] env[63088]: DEBUG oslo_vmware.api [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Task: {'id': task-1285357, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.270430] env[63088]: DEBUG oslo_vmware.api [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]5228baed-4728-d389-ec58-7a426ec3db90, 'name': SearchDatastore_Task, 'duration_secs': 0.015842} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.271720] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c78d1191-b09e-48e2-8be0-1f3f006180c7 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.277020] env[63088]: DEBUG oslo_vmware.api [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Waiting for the task: (returnval){ [ 985.277020] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52083b80-c71d-446e-b3a7-1f9e9ef6211e" [ 985.277020] env[63088]: _type = "Task" [ 985.277020] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.286511] env[63088]: DEBUG oslo_vmware.api [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52083b80-c71d-446e-b3a7-1f9e9ef6211e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.364583] env[63088]: DEBUG nova.compute.manager [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: c4fb31b2-1301-4a6e-bef2-8ea462caee0d] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 985.391385] env[63088]: DEBUG nova.virt.hardware [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 985.391667] env[63088]: DEBUG nova.virt.hardware [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 985.391827] env[63088]: DEBUG nova.virt.hardware [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 985.392023] env[63088]: DEBUG nova.virt.hardware [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 985.392180] env[63088]: DEBUG nova.virt.hardware [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 985.392336] env[63088]: DEBUG nova.virt.hardware [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 985.392644] env[63088]: DEBUG nova.virt.hardware [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 985.392821] env[63088]: DEBUG nova.virt.hardware [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 985.392993] env[63088]: DEBUG nova.virt.hardware [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 985.393175] env[63088]: DEBUG nova.virt.hardware [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 985.393352] env[63088]: DEBUG nova.virt.hardware [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 985.394225] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7f276b4-dc38-43db-bba5-1d11bb49a0ae {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.402612] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d28ace9-6182-446e-a262-33b817625a3a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.594494] env[63088]: DEBUG oslo_concurrency.lockutils [None req-82ce7dd7-1f78-4f20-808f-f5b9d77ac35d tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 985.618053] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bc161c1d-84e4-48a5-a6e5-b5d8ec5b62aa tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.274s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.621534] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 9.447s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 985.621534] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.621534] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63088) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 985.621534] env[63088]: DEBUG oslo_concurrency.lockutils [None req-56ee9ff4-78f5-42d5-895a-1b5958fb89c8 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.045s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 985.621534] env[63088]: DEBUG nova.objects.instance [None req-56ee9ff4-78f5-42d5-895a-1b5958fb89c8 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lazy-loading 'resources' on Instance uuid 66178bd0-7ddf-458c-98a2-3519e536dc63 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 985.627395] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a35b9d1-5eb0-4f55-aaa0-d322adaa03f3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.634037] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e0e794d-8e83-4821-b2df-ad64c60d4150 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.653026] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3050a986-abce-414e-9bd1-affae32cae06 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.661815] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-738d2e9c-a8df-423c-9529-22be7663252d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.692674] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180701MB free_disk=140GB free_vcpus=48 pci_devices=None {{(pid=63088) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 985.692842] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 985.706141] env[63088]: DEBUG oslo_vmware.api [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285356, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068606} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.706408] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] Extended root virtual disk {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 985.707199] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80fca059-45f7-4d9a-bbdc-932a0e776aea {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.741691] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] Reconfiguring VM instance instance-0000005d to attach disk [datastore1] 69645f50-46fc-4c15-9b39-1721f7636e31/69645f50-46fc-4c15-9b39-1721f7636e31.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 985.741691] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f8ca26d2-33eb-45bc-bda3-d42437877d33 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.756757] env[63088]: DEBUG oslo_vmware.api [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Task: {'id': task-1285357, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.758015] env[63088]: DEBUG oslo_vmware.api [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for the task: (returnval){ [ 985.758015] env[63088]: value = "task-1285358" [ 985.758015] env[63088]: _type = "Task" [ 985.758015] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.765936] env[63088]: DEBUG oslo_vmware.api [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285358, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.786318] env[63088]: DEBUG oslo_vmware.api [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52083b80-c71d-446e-b3a7-1f9e9ef6211e, 'name': SearchDatastore_Task, 'duration_secs': 0.009288} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.786580] env[63088]: DEBUG oslo_concurrency.lockutils [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 985.786845] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3/c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 985.787110] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6cb739ea-68f5-4653-98e3-001917ca255f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.793265] env[63088]: DEBUG oslo_vmware.api [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Waiting for the task: (returnval){ [ 985.793265] env[63088]: value = "task-1285359" [ 985.793265] env[63088]: _type = "Task" [ 985.793265] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.800625] env[63088]: DEBUG oslo_vmware.api [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Task: {'id': task-1285359, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.837363] env[63088]: DEBUG nova.compute.manager [req-8c95fe56-d7a0-4555-beac-3772107fb202 req-12234692-56d6-4eb8-8375-d4a36da35ef2 service nova] [instance: c4fb31b2-1301-4a6e-bef2-8ea462caee0d] Received event network-vif-plugged-88bc1e34-07ac-4819-9cd8-d933c1540bb2 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 985.837587] env[63088]: DEBUG oslo_concurrency.lockutils [req-8c95fe56-d7a0-4555-beac-3772107fb202 req-12234692-56d6-4eb8-8375-d4a36da35ef2 service nova] Acquiring lock "c4fb31b2-1301-4a6e-bef2-8ea462caee0d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 985.837796] env[63088]: DEBUG oslo_concurrency.lockutils [req-8c95fe56-d7a0-4555-beac-3772107fb202 req-12234692-56d6-4eb8-8375-d4a36da35ef2 service nova] Lock "c4fb31b2-1301-4a6e-bef2-8ea462caee0d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 985.838023] env[63088]: DEBUG oslo_concurrency.lockutils [req-8c95fe56-d7a0-4555-beac-3772107fb202 req-12234692-56d6-4eb8-8375-d4a36da35ef2 service nova] Lock "c4fb31b2-1301-4a6e-bef2-8ea462caee0d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.838161] env[63088]: DEBUG nova.compute.manager [req-8c95fe56-d7a0-4555-beac-3772107fb202 req-12234692-56d6-4eb8-8375-d4a36da35ef2 service nova] [instance: c4fb31b2-1301-4a6e-bef2-8ea462caee0d] No waiting events found dispatching network-vif-plugged-88bc1e34-07ac-4819-9cd8-d933c1540bb2 {{(pid=63088) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 985.838334] env[63088]: WARNING nova.compute.manager [req-8c95fe56-d7a0-4555-beac-3772107fb202 req-12234692-56d6-4eb8-8375-d4a36da35ef2 service nova] [instance: c4fb31b2-1301-4a6e-bef2-8ea462caee0d] Received unexpected event network-vif-plugged-88bc1e34-07ac-4819-9cd8-d933c1540bb2 for instance with vm_state building and task_state spawning. [ 986.033921] env[63088]: DEBUG nova.network.neutron [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: c4fb31b2-1301-4a6e-bef2-8ea462caee0d] Successfully updated port: 88bc1e34-07ac-4819-9cd8-d933c1540bb2 {{(pid=63088) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 986.101163] env[63088]: DEBUG oslo_concurrency.lockutils [None req-784c107b-9a9f-49f2-ad33-32d44d1bf7ff tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquiring lock "refresh_cache-f582da15-dfc3-45e1-a995-9dd0c9533869" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 986.101385] env[63088]: DEBUG oslo_concurrency.lockutils [None req-784c107b-9a9f-49f2-ad33-32d44d1bf7ff tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquired lock "refresh_cache-f582da15-dfc3-45e1-a995-9dd0c9533869" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 986.101581] env[63088]: DEBUG nova.network.neutron [None req-784c107b-9a9f-49f2-ad33-32d44d1bf7ff tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 986.197992] env[63088]: INFO nova.scheduler.client.report [None req-bc161c1d-84e4-48a5-a6e5-b5d8ec5b62aa tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Deleted allocation for migration 391d39ab-da7f-476e-bf0e-659212e03687 [ 986.237847] env[63088]: DEBUG oslo_vmware.api [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Task: {'id': task-1285357, 'name': PowerOnVM_Task, 'duration_secs': 1.013891} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.238210] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: 13388466-4adb-4d56-9fc1-e3f5dc516077] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 986.240575] env[63088]: INFO nova.compute.manager [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: 13388466-4adb-4d56-9fc1-e3f5dc516077] Took 5.38 seconds to spawn the instance on the hypervisor. [ 986.240575] env[63088]: DEBUG nova.compute.manager [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: 13388466-4adb-4d56-9fc1-e3f5dc516077] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 986.240575] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d61fd039-f91d-4652-88eb-904d2098ffdd {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.273370] env[63088]: DEBUG oslo_vmware.api [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285358, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.306422] env[63088]: DEBUG oslo_vmware.api [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Task: {'id': task-1285359, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.320583] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-250766a4-d345-4c67-a6a2-97d21404b16f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.329070] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-450cff22-31fa-4a5f-ac80-84c7deaf3ef4 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.361251] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff3ba48f-d5e0-4957-a8de-402da3f121c8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.370028] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8db71092-b17a-409e-9f2b-d51163bceac1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.386504] env[63088]: DEBUG nova.compute.provider_tree [None req-56ee9ff4-78f5-42d5-895a-1b5958fb89c8 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 986.540359] env[63088]: DEBUG oslo_concurrency.lockutils [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquiring lock "refresh_cache-c4fb31b2-1301-4a6e-bef2-8ea462caee0d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 986.540943] env[63088]: DEBUG oslo_concurrency.lockutils [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquired lock "refresh_cache-c4fb31b2-1301-4a6e-bef2-8ea462caee0d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 986.540943] env[63088]: DEBUG nova.network.neutron [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: c4fb31b2-1301-4a6e-bef2-8ea462caee0d] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 986.707606] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bc161c1d-84e4-48a5-a6e5-b5d8ec5b62aa tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Lock "2ad33254-8030-4454-b023-3359071f85de" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 16.659s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.758396] env[63088]: INFO nova.compute.manager [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: 13388466-4adb-4d56-9fc1-e3f5dc516077] Took 14.41 seconds to build instance. [ 986.771486] env[63088]: DEBUG oslo_vmware.api [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285358, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.807634] env[63088]: DEBUG oslo_vmware.api [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Task: {'id': task-1285359, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.94814} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.810134] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3/c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 986.810452] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3] Extending root virtual disk to 1048576 {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 986.810739] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-64ed0e12-6b56-4fdf-b46a-6a52661bbe77 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.819479] env[63088]: DEBUG oslo_vmware.api [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Waiting for the task: (returnval){ [ 986.819479] env[63088]: value = "task-1285360" [ 986.819479] env[63088]: _type = "Task" [ 986.819479] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.829061] env[63088]: DEBUG oslo_vmware.api [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Task: {'id': task-1285360, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.853294] env[63088]: INFO nova.network.neutron [None req-784c107b-9a9f-49f2-ad33-32d44d1bf7ff tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Port 051455f3-4c94-4050-8610-bd8ace9f3988 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 986.853702] env[63088]: DEBUG nova.network.neutron [None req-784c107b-9a9f-49f2-ad33-32d44d1bf7ff tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Updating instance_info_cache with network_info: [{"id": "81fc2dda-617e-4e4c-aa91-7e7020003995", "address": "fa:16:3e:8c:09:4d", "network": {"id": "1174a284-d3ca-4f9e-aa81-13ee9a693e55", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1994276040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa3d24a1a6c0430985fd80365d986ee1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap81fc2dda-61", "ovs_interfaceid": "81fc2dda-617e-4e4c-aa91-7e7020003995", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 986.889806] env[63088]: DEBUG nova.scheduler.client.report [None req-56ee9ff4-78f5-42d5-895a-1b5958fb89c8 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 986.956793] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c61aefb1-d0bc-40a0-9a74-b1d68160c9fd tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquiring lock "interface-d6fef3ef-fec8-4929-b9b8-5e63306aa51d-051455f3-4c94-4050-8610-bd8ace9f3988" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.957112] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c61aefb1-d0bc-40a0-9a74-b1d68160c9fd tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Lock "interface-d6fef3ef-fec8-4929-b9b8-5e63306aa51d-051455f3-4c94-4050-8610-bd8ace9f3988" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.957654] env[63088]: DEBUG nova.objects.instance [None req-c61aefb1-d0bc-40a0-9a74-b1d68160c9fd tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Lazy-loading 'flavor' on Instance uuid d6fef3ef-fec8-4929-b9b8-5e63306aa51d {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 987.076443] env[63088]: DEBUG nova.network.neutron [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: c4fb31b2-1301-4a6e-bef2-8ea462caee0d] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 987.145456] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0de4c45b-7355-4582-b6ec-c78d8eef867d tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquiring lock "2ad33254-8030-4454-b023-3359071f85de" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 987.145708] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0de4c45b-7355-4582-b6ec-c78d8eef867d tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Lock "2ad33254-8030-4454-b023-3359071f85de" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 987.145926] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0de4c45b-7355-4582-b6ec-c78d8eef867d tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquiring lock "2ad33254-8030-4454-b023-3359071f85de-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 987.146129] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0de4c45b-7355-4582-b6ec-c78d8eef867d tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Lock "2ad33254-8030-4454-b023-3359071f85de-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 987.146330] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0de4c45b-7355-4582-b6ec-c78d8eef867d tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Lock "2ad33254-8030-4454-b023-3359071f85de-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 987.148430] env[63088]: INFO nova.compute.manager [None req-0de4c45b-7355-4582-b6ec-c78d8eef867d tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2ad33254-8030-4454-b023-3359071f85de] Terminating instance [ 987.150228] env[63088]: DEBUG nova.compute.manager [None req-0de4c45b-7355-4582-b6ec-c78d8eef867d tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2ad33254-8030-4454-b023-3359071f85de] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 987.150404] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-0de4c45b-7355-4582-b6ec-c78d8eef867d tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2ad33254-8030-4454-b023-3359071f85de] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 987.151254] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1584438-3cd0-46ed-8d19-a05a6918e58d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.159670] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-0de4c45b-7355-4582-b6ec-c78d8eef867d tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2ad33254-8030-4454-b023-3359071f85de] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 987.159772] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d01ed3cb-67d9-45eb-b1d3-b833b7c9386b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.166789] env[63088]: DEBUG oslo_vmware.api [None req-0de4c45b-7355-4582-b6ec-c78d8eef867d tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Waiting for the task: (returnval){ [ 987.166789] env[63088]: value = "task-1285361" [ 987.166789] env[63088]: _type = "Task" [ 987.166789] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.176946] env[63088]: DEBUG oslo_vmware.api [None req-0de4c45b-7355-4582-b6ec-c78d8eef867d tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285361, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.220048] env[63088]: DEBUG nova.network.neutron [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: c4fb31b2-1301-4a6e-bef2-8ea462caee0d] Updating instance_info_cache with network_info: [{"id": "88bc1e34-07ac-4819-9cd8-d933c1540bb2", "address": "fa:16:3e:94:84:df", "network": {"id": "307b966a-d9e1-40fd-9313-1ad94c734308", "bridge": "br-int", "label": "tempest-ServersTestJSON-1907249197-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "03e20d606e654362acbe2b36fe499ae3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae18b41f-e73c-44f1-83dd-467c080944f4", "external-id": "nsx-vlan-transportzone-653", "segmentation_id": 653, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap88bc1e34-07", "ovs_interfaceid": "88bc1e34-07ac-4819-9cd8-d933c1540bb2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 987.260815] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5d171fc5-802c-4f86-8108-e3679e614b9b tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Lock "13388466-4adb-4d56-9fc1-e3f5dc516077" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.920s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 987.271962] env[63088]: DEBUG oslo_vmware.api [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285358, 'name': ReconfigVM_Task, 'duration_secs': 1.191545} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.272195] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] Reconfigured VM instance instance-0000005d to attach disk [datastore1] 69645f50-46fc-4c15-9b39-1721f7636e31/69645f50-46fc-4c15-9b39-1721f7636e31.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 987.272827] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f970b825-2a13-4980-a60e-4d09c0df7977 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.279870] env[63088]: DEBUG oslo_vmware.api [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for the task: (returnval){ [ 987.279870] env[63088]: value = "task-1285362" [ 987.279870] env[63088]: _type = "Task" [ 987.279870] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.289460] env[63088]: DEBUG oslo_vmware.api [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285362, 'name': Rename_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.332631] env[63088]: DEBUG oslo_vmware.api [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Task: {'id': task-1285360, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.161494} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.333042] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3] Extended root virtual disk {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 987.333890] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f78dd92-3855-4940-bce0-eedf1acef972 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.355399] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3] Reconfiguring VM instance instance-0000005f to attach disk [datastore1] c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3/c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 987.355979] env[63088]: DEBUG oslo_concurrency.lockutils [None req-784c107b-9a9f-49f2-ad33-32d44d1bf7ff tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Releasing lock "refresh_cache-f582da15-dfc3-45e1-a995-9dd0c9533869" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 987.358037] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-11fe41c4-d1b2-4b90-9cf0-c94d211c24f4 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.380712] env[63088]: DEBUG oslo_vmware.api [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Waiting for the task: (returnval){ [ 987.380712] env[63088]: value = "task-1285363" [ 987.380712] env[63088]: _type = "Task" [ 987.380712] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.390636] env[63088]: DEBUG oslo_vmware.api [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Task: {'id': task-1285363, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.395354] env[63088]: DEBUG oslo_concurrency.lockutils [None req-56ee9ff4-78f5-42d5-895a-1b5958fb89c8 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.774s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 987.397586] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.356s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 987.397825] env[63088]: DEBUG nova.objects.instance [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Lazy-loading 'pci_requests' on Instance uuid 71e37d8e-a454-46c4-a3cc-3d5671a32beb {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 987.413260] env[63088]: INFO nova.scheduler.client.report [None req-56ee9ff4-78f5-42d5-895a-1b5958fb89c8 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Deleted allocations for instance 66178bd0-7ddf-458c-98a2-3519e536dc63 [ 987.560126] env[63088]: DEBUG nova.objects.instance [None req-c61aefb1-d0bc-40a0-9a74-b1d68160c9fd tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Lazy-loading 'pci_requests' on Instance uuid d6fef3ef-fec8-4929-b9b8-5e63306aa51d {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 987.625054] env[63088]: DEBUG oslo_concurrency.lockutils [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquiring lock "a39840c1-3d60-478b-987c-2519551963cf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 987.625310] env[63088]: DEBUG oslo_concurrency.lockutils [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Lock "a39840c1-3d60-478b-987c-2519551963cf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 987.677168] env[63088]: DEBUG oslo_vmware.api [None req-0de4c45b-7355-4582-b6ec-c78d8eef867d tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285361, 'name': PowerOffVM_Task, 'duration_secs': 0.191763} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.677372] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-0de4c45b-7355-4582-b6ec-c78d8eef867d tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2ad33254-8030-4454-b023-3359071f85de] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 987.677546] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-0de4c45b-7355-4582-b6ec-c78d8eef867d tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2ad33254-8030-4454-b023-3359071f85de] Unregistering the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 987.677807] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a7ccd798-24e6-4f29-9630-a06af4b10cb9 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.722426] env[63088]: DEBUG oslo_concurrency.lockutils [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Releasing lock "refresh_cache-c4fb31b2-1301-4a6e-bef2-8ea462caee0d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 987.722794] env[63088]: DEBUG nova.compute.manager [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: c4fb31b2-1301-4a6e-bef2-8ea462caee0d] Instance network_info: |[{"id": "88bc1e34-07ac-4819-9cd8-d933c1540bb2", "address": "fa:16:3e:94:84:df", "network": {"id": "307b966a-d9e1-40fd-9313-1ad94c734308", "bridge": "br-int", "label": "tempest-ServersTestJSON-1907249197-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "03e20d606e654362acbe2b36fe499ae3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae18b41f-e73c-44f1-83dd-467c080944f4", "external-id": "nsx-vlan-transportzone-653", "segmentation_id": 653, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap88bc1e34-07", "ovs_interfaceid": "88bc1e34-07ac-4819-9cd8-d933c1540bb2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 987.723254] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: c4fb31b2-1301-4a6e-bef2-8ea462caee0d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:94:84:df', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ae18b41f-e73c-44f1-83dd-467c080944f4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '88bc1e34-07ac-4819-9cd8-d933c1540bb2', 'vif_model': 'vmxnet3'}] {{(pid=63088) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 987.731088] env[63088]: DEBUG oslo.service.loopingcall [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 987.731321] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c4fb31b2-1301-4a6e-bef2-8ea462caee0d] Creating VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 987.731580] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9e23f380-399c-44ce-8a5d-9ca995b79362 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.752781] env[63088]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 987.752781] env[63088]: value = "task-1285365" [ 987.752781] env[63088]: _type = "Task" [ 987.752781] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.761034] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-0de4c45b-7355-4582-b6ec-c78d8eef867d tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2ad33254-8030-4454-b023-3359071f85de] Unregistered the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 987.761204] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-0de4c45b-7355-4582-b6ec-c78d8eef867d tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2ad33254-8030-4454-b023-3359071f85de] Deleting contents of the VM from datastore datastore2 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 987.761497] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-0de4c45b-7355-4582-b6ec-c78d8eef867d tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Deleting the datastore file [datastore2] 2ad33254-8030-4454-b023-3359071f85de {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 987.761762] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-efe46fa9-7613-4b56-a7b9-a88198c07e56 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.768337] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285365, 'name': CreateVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.774290] env[63088]: DEBUG oslo_vmware.api [None req-0de4c45b-7355-4582-b6ec-c78d8eef867d tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Waiting for the task: (returnval){ [ 987.774290] env[63088]: value = "task-1285366" [ 987.774290] env[63088]: _type = "Task" [ 987.774290] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.782373] env[63088]: DEBUG oslo_vmware.api [None req-0de4c45b-7355-4582-b6ec-c78d8eef867d tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285366, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.790480] env[63088]: DEBUG oslo_vmware.api [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285362, 'name': Rename_Task, 'duration_secs': 0.17798} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.790770] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 987.791031] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-32d183c2-c064-4a36-8f1e-6f86739277c9 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.797383] env[63088]: DEBUG oslo_vmware.api [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for the task: (returnval){ [ 987.797383] env[63088]: value = "task-1285367" [ 987.797383] env[63088]: _type = "Task" [ 987.797383] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.805815] env[63088]: DEBUG oslo_vmware.api [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285367, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.875368] env[63088]: DEBUG oslo_concurrency.lockutils [None req-784c107b-9a9f-49f2-ad33-32d44d1bf7ff tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Lock "interface-f582da15-dfc3-45e1-a995-9dd0c9533869-051455f3-4c94-4050-8610-bd8ace9f3988" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.803s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 987.895687] env[63088]: DEBUG oslo_vmware.api [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Task: {'id': task-1285363, 'name': ReconfigVM_Task, 'duration_secs': 0.276227} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.896052] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3] Reconfigured VM instance instance-0000005f to attach disk [datastore1] c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3/c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 987.896702] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-da438d64-b522-4dff-ab6f-e1bce5e14a6c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.902874] env[63088]: DEBUG nova.objects.instance [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Lazy-loading 'numa_topology' on Instance uuid 71e37d8e-a454-46c4-a3cc-3d5671a32beb {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 987.906116] env[63088]: DEBUG oslo_vmware.api [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Waiting for the task: (returnval){ [ 987.906116] env[63088]: value = "task-1285368" [ 987.906116] env[63088]: _type = "Task" [ 987.906116] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.916724] env[63088]: DEBUG oslo_vmware.api [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Task: {'id': task-1285368, 'name': Rename_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.921718] env[63088]: DEBUG oslo_concurrency.lockutils [None req-56ee9ff4-78f5-42d5-895a-1b5958fb89c8 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "66178bd0-7ddf-458c-98a2-3519e536dc63" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.749s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.055190] env[63088]: DEBUG nova.compute.manager [req-4a342f9b-676f-4f0c-a011-bd7d57b2fe6d req-68990000-f6c1-47ba-b896-9c4076b804b5 service nova] [instance: c4fb31b2-1301-4a6e-bef2-8ea462caee0d] Received event network-changed-88bc1e34-07ac-4819-9cd8-d933c1540bb2 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 988.055371] env[63088]: DEBUG nova.compute.manager [req-4a342f9b-676f-4f0c-a011-bd7d57b2fe6d req-68990000-f6c1-47ba-b896-9c4076b804b5 service nova] [instance: c4fb31b2-1301-4a6e-bef2-8ea462caee0d] Refreshing instance network info cache due to event network-changed-88bc1e34-07ac-4819-9cd8-d933c1540bb2. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 988.055720] env[63088]: DEBUG oslo_concurrency.lockutils [req-4a342f9b-676f-4f0c-a011-bd7d57b2fe6d req-68990000-f6c1-47ba-b896-9c4076b804b5 service nova] Acquiring lock "refresh_cache-c4fb31b2-1301-4a6e-bef2-8ea462caee0d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 988.055866] env[63088]: DEBUG oslo_concurrency.lockutils [req-4a342f9b-676f-4f0c-a011-bd7d57b2fe6d req-68990000-f6c1-47ba-b896-9c4076b804b5 service nova] Acquired lock "refresh_cache-c4fb31b2-1301-4a6e-bef2-8ea462caee0d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.056096] env[63088]: DEBUG nova.network.neutron [req-4a342f9b-676f-4f0c-a011-bd7d57b2fe6d req-68990000-f6c1-47ba-b896-9c4076b804b5 service nova] [instance: c4fb31b2-1301-4a6e-bef2-8ea462caee0d] Refreshing network info cache for port 88bc1e34-07ac-4819-9cd8-d933c1540bb2 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 988.065017] env[63088]: DEBUG nova.objects.base [None req-c61aefb1-d0bc-40a0-9a74-b1d68160c9fd tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=63088) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 988.065250] env[63088]: DEBUG nova.network.neutron [None req-c61aefb1-d0bc-40a0-9a74-b1d68160c9fd tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 988.128033] env[63088]: DEBUG nova.compute.manager [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: a39840c1-3d60-478b-987c-2519551963cf] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 988.151537] env[63088]: DEBUG nova.policy [None req-c61aefb1-d0bc-40a0-9a74-b1d68160c9fd tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '93f1d334e71a4a99ba300c3ee6d70243', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'aa3d24a1a6c0430985fd80365d986ee1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 988.268839] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285365, 'name': CreateVM_Task, 'duration_secs': 0.344176} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.273022] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c4fb31b2-1301-4a6e-bef2-8ea462caee0d] Created VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 988.273022] env[63088]: DEBUG oslo_concurrency.lockutils [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 988.273022] env[63088]: DEBUG oslo_concurrency.lockutils [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.273022] env[63088]: DEBUG oslo_concurrency.lockutils [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 988.273022] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2a0ba3f8-b8d4-4f42-b381-c55449b11ba2 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.279730] env[63088]: DEBUG oslo_vmware.api [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Waiting for the task: (returnval){ [ 988.279730] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]5236da32-f0e7-9795-7fb4-2382639ed45a" [ 988.279730] env[63088]: _type = "Task" [ 988.279730] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.287331] env[63088]: DEBUG oslo_vmware.api [None req-0de4c45b-7355-4582-b6ec-c78d8eef867d tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285366, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.213655} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.288174] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-0de4c45b-7355-4582-b6ec-c78d8eef867d tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 988.288524] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-0de4c45b-7355-4582-b6ec-c78d8eef867d tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2ad33254-8030-4454-b023-3359071f85de] Deleted contents of the VM from datastore datastore2 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 988.288840] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-0de4c45b-7355-4582-b6ec-c78d8eef867d tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2ad33254-8030-4454-b023-3359071f85de] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 988.289210] env[63088]: INFO nova.compute.manager [None req-0de4c45b-7355-4582-b6ec-c78d8eef867d tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: 2ad33254-8030-4454-b023-3359071f85de] Took 1.14 seconds to destroy the instance on the hypervisor. [ 988.289649] env[63088]: DEBUG oslo.service.loopingcall [None req-0de4c45b-7355-4582-b6ec-c78d8eef867d tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 988.290091] env[63088]: DEBUG nova.compute.manager [-] [instance: 2ad33254-8030-4454-b023-3359071f85de] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 988.290435] env[63088]: DEBUG nova.network.neutron [-] [instance: 2ad33254-8030-4454-b023-3359071f85de] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 988.296101] env[63088]: DEBUG oslo_vmware.api [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]5236da32-f0e7-9795-7fb4-2382639ed45a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.307713] env[63088]: DEBUG oslo_vmware.api [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285367, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.406721] env[63088]: INFO nova.compute.claims [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 988.419829] env[63088]: DEBUG oslo_vmware.api [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Task: {'id': task-1285368, 'name': Rename_Task, 'duration_secs': 0.183212} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.420137] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 988.420432] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-55ece065-e8d1-46ff-9ff4-ba8e48de5719 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.429259] env[63088]: DEBUG oslo_vmware.api [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Waiting for the task: (returnval){ [ 988.429259] env[63088]: value = "task-1285369" [ 988.429259] env[63088]: _type = "Task" [ 988.429259] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.439232] env[63088]: DEBUG oslo_vmware.api [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Task: {'id': task-1285369, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.492737] env[63088]: DEBUG oslo_concurrency.lockutils [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquiring lock "32418785-8823-4a5e-90a9-e205752530b7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 988.492987] env[63088]: DEBUG oslo_concurrency.lockutils [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "32418785-8823-4a5e-90a9-e205752530b7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 988.650097] env[63088]: DEBUG oslo_concurrency.lockutils [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 988.713441] env[63088]: DEBUG nova.compute.manager [req-6011d62d-8eb4-4550-b528-1413ed120111 req-e8cec2c8-d1ec-4cb0-8dfd-7d9ef58185c0 service nova] [instance: 2ad33254-8030-4454-b023-3359071f85de] Received event network-vif-deleted-1d2bee7e-bbc7-4b17-976b-47e7bbb268af {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 988.713687] env[63088]: INFO nova.compute.manager [req-6011d62d-8eb4-4550-b528-1413ed120111 req-e8cec2c8-d1ec-4cb0-8dfd-7d9ef58185c0 service nova] [instance: 2ad33254-8030-4454-b023-3359071f85de] Neutron deleted interface 1d2bee7e-bbc7-4b17-976b-47e7bbb268af; detaching it from the instance and deleting it from the info cache [ 988.713889] env[63088]: DEBUG nova.network.neutron [req-6011d62d-8eb4-4550-b528-1413ed120111 req-e8cec2c8-d1ec-4cb0-8dfd-7d9ef58185c0 service nova] [instance: 2ad33254-8030-4454-b023-3359071f85de] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 988.793698] env[63088]: DEBUG oslo_vmware.api [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]5236da32-f0e7-9795-7fb4-2382639ed45a, 'name': SearchDatastore_Task, 'duration_secs': 0.016166} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.793989] env[63088]: DEBUG oslo_concurrency.lockutils [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 988.794260] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: c4fb31b2-1301-4a6e-bef2-8ea462caee0d] Processing image 1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 988.794527] env[63088]: DEBUG oslo_concurrency.lockutils [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 988.794721] env[63088]: DEBUG oslo_concurrency.lockutils [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.794945] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 988.795241] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-53121979-7ff8-4434-835f-feae4f6eeee7 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.810079] env[63088]: DEBUG oslo_vmware.api [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285367, 'name': PowerOnVM_Task, 'duration_secs': 0.528199} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.811161] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 988.811440] env[63088]: INFO nova.compute.manager [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] Took 10.26 seconds to spawn the instance on the hypervisor. [ 988.811667] env[63088]: DEBUG nova.compute.manager [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 988.812018] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 988.812198] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 988.813341] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3baaddd-b01d-4d53-a64a-3e0b09a84a61 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.815797] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ddc1ba89-acc4-4b8d-8576-619ab74a67c1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.821424] env[63088]: DEBUG oslo_vmware.api [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Waiting for the task: (returnval){ [ 988.821424] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]527597cc-e362-41f8-45b0-a4f19191faf1" [ 988.821424] env[63088]: _type = "Task" [ 988.821424] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.840717] env[63088]: DEBUG oslo_vmware.api [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]527597cc-e362-41f8-45b0-a4f19191faf1, 'name': SearchDatastore_Task, 'duration_secs': 0.011406} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.841708] env[63088]: DEBUG nova.network.neutron [req-4a342f9b-676f-4f0c-a011-bd7d57b2fe6d req-68990000-f6c1-47ba-b896-9c4076b804b5 service nova] [instance: c4fb31b2-1301-4a6e-bef2-8ea462caee0d] Updated VIF entry in instance network info cache for port 88bc1e34-07ac-4819-9cd8-d933c1540bb2. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 988.842062] env[63088]: DEBUG nova.network.neutron [req-4a342f9b-676f-4f0c-a011-bd7d57b2fe6d req-68990000-f6c1-47ba-b896-9c4076b804b5 service nova] [instance: c4fb31b2-1301-4a6e-bef2-8ea462caee0d] Updating instance_info_cache with network_info: [{"id": "88bc1e34-07ac-4819-9cd8-d933c1540bb2", "address": "fa:16:3e:94:84:df", "network": {"id": "307b966a-d9e1-40fd-9313-1ad94c734308", "bridge": "br-int", "label": "tempest-ServersTestJSON-1907249197-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "03e20d606e654362acbe2b36fe499ae3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae18b41f-e73c-44f1-83dd-467c080944f4", "external-id": "nsx-vlan-transportzone-653", "segmentation_id": 653, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap88bc1e34-07", "ovs_interfaceid": "88bc1e34-07ac-4819-9cd8-d933c1540bb2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 988.844073] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2d948546-bdca-42d5-9534-e399a555faa1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.851045] env[63088]: DEBUG oslo_vmware.api [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Waiting for the task: (returnval){ [ 988.851045] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52f24dd6-0c78-15c0-8633-c3fbc50d8237" [ 988.851045] env[63088]: _type = "Task" [ 988.851045] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.861204] env[63088]: DEBUG oslo_vmware.api [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52f24dd6-0c78-15c0-8633-c3fbc50d8237, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.942141] env[63088]: DEBUG oslo_vmware.api [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Task: {'id': task-1285369, 'name': PowerOnVM_Task} progress is 90%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.996867] env[63088]: DEBUG nova.compute.manager [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 32418785-8823-4a5e-90a9-e205752530b7] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 989.044279] env[63088]: DEBUG nova.network.neutron [-] [instance: 2ad33254-8030-4454-b023-3359071f85de] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 989.216902] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5a3594a4-7c69-4ce8-8f09-8dba2451f905 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.227594] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2cccf13-01ba-4e67-9499-2db07f4c7040 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.274321] env[63088]: DEBUG nova.compute.manager [req-6011d62d-8eb4-4550-b528-1413ed120111 req-e8cec2c8-d1ec-4cb0-8dfd-7d9ef58185c0 service nova] [instance: 2ad33254-8030-4454-b023-3359071f85de] Detach interface failed, port_id=1d2bee7e-bbc7-4b17-976b-47e7bbb268af, reason: Instance 2ad33254-8030-4454-b023-3359071f85de could not be found. {{(pid=63088) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 989.341660] env[63088]: INFO nova.compute.manager [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] Took 19.80 seconds to build instance. [ 989.344893] env[63088]: DEBUG oslo_concurrency.lockutils [req-4a342f9b-676f-4f0c-a011-bd7d57b2fe6d req-68990000-f6c1-47ba-b896-9c4076b804b5 service nova] Releasing lock "refresh_cache-c4fb31b2-1301-4a6e-bef2-8ea462caee0d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 989.345156] env[63088]: DEBUG nova.compute.manager [req-4a342f9b-676f-4f0c-a011-bd7d57b2fe6d req-68990000-f6c1-47ba-b896-9c4076b804b5 service nova] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Received event network-changed-81fc2dda-617e-4e4c-aa91-7e7020003995 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 989.345331] env[63088]: DEBUG nova.compute.manager [req-4a342f9b-676f-4f0c-a011-bd7d57b2fe6d req-68990000-f6c1-47ba-b896-9c4076b804b5 service nova] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Refreshing instance network info cache due to event network-changed-81fc2dda-617e-4e4c-aa91-7e7020003995. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 989.345538] env[63088]: DEBUG oslo_concurrency.lockutils [req-4a342f9b-676f-4f0c-a011-bd7d57b2fe6d req-68990000-f6c1-47ba-b896-9c4076b804b5 service nova] Acquiring lock "refresh_cache-f582da15-dfc3-45e1-a995-9dd0c9533869" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 989.345681] env[63088]: DEBUG oslo_concurrency.lockutils [req-4a342f9b-676f-4f0c-a011-bd7d57b2fe6d req-68990000-f6c1-47ba-b896-9c4076b804b5 service nova] Acquired lock "refresh_cache-f582da15-dfc3-45e1-a995-9dd0c9533869" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.345843] env[63088]: DEBUG nova.network.neutron [req-4a342f9b-676f-4f0c-a011-bd7d57b2fe6d req-68990000-f6c1-47ba-b896-9c4076b804b5 service nova] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Refreshing network info cache for port 81fc2dda-617e-4e4c-aa91-7e7020003995 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 989.362722] env[63088]: DEBUG oslo_vmware.api [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52f24dd6-0c78-15c0-8633-c3fbc50d8237, 'name': SearchDatastore_Task, 'duration_secs': 0.011377} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.362984] env[63088]: DEBUG oslo_concurrency.lockutils [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 989.363250] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] c4fb31b2-1301-4a6e-bef2-8ea462caee0d/c4fb31b2-1301-4a6e-bef2-8ea462caee0d.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 989.363505] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9b6d1fad-b922-4ea1-85a4-c517f475fa18 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.371504] env[63088]: DEBUG oslo_vmware.api [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Waiting for the task: (returnval){ [ 989.371504] env[63088]: value = "task-1285370" [ 989.371504] env[63088]: _type = "Task" [ 989.371504] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.379820] env[63088]: DEBUG oslo_vmware.api [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285370, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.442144] env[63088]: DEBUG oslo_vmware.api [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Task: {'id': task-1285369, 'name': PowerOnVM_Task, 'duration_secs': 0.596536} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.442144] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 989.442144] env[63088]: INFO nova.compute.manager [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3] Took 6.31 seconds to spawn the instance on the hypervisor. [ 989.442440] env[63088]: DEBUG nova.compute.manager [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 989.443269] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf60d9f7-da2d-43ee-ad1b-d388392c0990 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.519221] env[63088]: DEBUG oslo_concurrency.lockutils [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 989.546831] env[63088]: INFO nova.compute.manager [-] [instance: 2ad33254-8030-4454-b023-3359071f85de] Took 1.26 seconds to deallocate network for instance. [ 989.586355] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee05bb31-4aeb-49b1-a806-feff8de91667 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.598270] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e56f75d-c28d-4499-a73a-53419931a4ab {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.640035] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc8bf293-79f8-4de3-ab4a-0b0d8b3d66d5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.649671] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e0d0a02-efc2-42b2-90ee-241b469128c2 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.663899] env[63088]: DEBUG nova.compute.provider_tree [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 989.838666] env[63088]: DEBUG nova.network.neutron [None req-c61aefb1-d0bc-40a0-9a74-b1d68160c9fd tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Successfully updated port: 051455f3-4c94-4050-8610-bd8ace9f3988 {{(pid=63088) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 989.843586] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6a589f09-ee34-4a17-979d-f4715dda4b5a tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lock "69645f50-46fc-4c15-9b39-1721f7636e31" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.315s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.881789] env[63088]: DEBUG oslo_vmware.api [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285370, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.962385] env[63088]: INFO nova.compute.manager [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3] Took 17.16 seconds to build instance. [ 990.053782] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0de4c45b-7355-4582-b6ec-c78d8eef867d tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 990.066358] env[63088]: DEBUG nova.network.neutron [req-4a342f9b-676f-4f0c-a011-bd7d57b2fe6d req-68990000-f6c1-47ba-b896-9c4076b804b5 service nova] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Updated VIF entry in instance network info cache for port 81fc2dda-617e-4e4c-aa91-7e7020003995. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 990.066741] env[63088]: DEBUG nova.network.neutron [req-4a342f9b-676f-4f0c-a011-bd7d57b2fe6d req-68990000-f6c1-47ba-b896-9c4076b804b5 service nova] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Updating instance_info_cache with network_info: [{"id": "81fc2dda-617e-4e4c-aa91-7e7020003995", "address": "fa:16:3e:8c:09:4d", "network": {"id": "1174a284-d3ca-4f9e-aa81-13ee9a693e55", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1994276040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa3d24a1a6c0430985fd80365d986ee1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap81fc2dda-61", "ovs_interfaceid": "81fc2dda-617e-4e4c-aa91-7e7020003995", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 990.166843] env[63088]: DEBUG nova.scheduler.client.report [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 990.244440] env[63088]: DEBUG nova.compute.manager [req-591139ea-665b-41fa-8be8-e3cbccba845d req-4286506b-656c-4767-9814-a9ec579d6403 service nova] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Received event network-vif-plugged-051455f3-4c94-4050-8610-bd8ace9f3988 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 990.244662] env[63088]: DEBUG oslo_concurrency.lockutils [req-591139ea-665b-41fa-8be8-e3cbccba845d req-4286506b-656c-4767-9814-a9ec579d6403 service nova] Acquiring lock "d6fef3ef-fec8-4929-b9b8-5e63306aa51d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 990.244931] env[63088]: DEBUG oslo_concurrency.lockutils [req-591139ea-665b-41fa-8be8-e3cbccba845d req-4286506b-656c-4767-9814-a9ec579d6403 service nova] Lock "d6fef3ef-fec8-4929-b9b8-5e63306aa51d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 990.245518] env[63088]: DEBUG oslo_concurrency.lockutils [req-591139ea-665b-41fa-8be8-e3cbccba845d req-4286506b-656c-4767-9814-a9ec579d6403 service nova] Lock "d6fef3ef-fec8-4929-b9b8-5e63306aa51d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 990.245842] env[63088]: DEBUG nova.compute.manager [req-591139ea-665b-41fa-8be8-e3cbccba845d req-4286506b-656c-4767-9814-a9ec579d6403 service nova] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] No waiting events found dispatching network-vif-plugged-051455f3-4c94-4050-8610-bd8ace9f3988 {{(pid=63088) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 990.246209] env[63088]: WARNING nova.compute.manager [req-591139ea-665b-41fa-8be8-e3cbccba845d req-4286506b-656c-4767-9814-a9ec579d6403 service nova] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Received unexpected event network-vif-plugged-051455f3-4c94-4050-8610-bd8ace9f3988 for instance with vm_state active and task_state None. [ 990.246643] env[63088]: DEBUG nova.compute.manager [req-591139ea-665b-41fa-8be8-e3cbccba845d req-4286506b-656c-4767-9814-a9ec579d6403 service nova] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Received event network-changed-051455f3-4c94-4050-8610-bd8ace9f3988 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 990.246984] env[63088]: DEBUG nova.compute.manager [req-591139ea-665b-41fa-8be8-e3cbccba845d req-4286506b-656c-4767-9814-a9ec579d6403 service nova] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Refreshing instance network info cache due to event network-changed-051455f3-4c94-4050-8610-bd8ace9f3988. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 990.247437] env[63088]: DEBUG oslo_concurrency.lockutils [req-591139ea-665b-41fa-8be8-e3cbccba845d req-4286506b-656c-4767-9814-a9ec579d6403 service nova] Acquiring lock "refresh_cache-d6fef3ef-fec8-4929-b9b8-5e63306aa51d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 990.247652] env[63088]: DEBUG oslo_concurrency.lockutils [req-591139ea-665b-41fa-8be8-e3cbccba845d req-4286506b-656c-4767-9814-a9ec579d6403 service nova] Acquired lock "refresh_cache-d6fef3ef-fec8-4929-b9b8-5e63306aa51d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.247937] env[63088]: DEBUG nova.network.neutron [req-591139ea-665b-41fa-8be8-e3cbccba845d req-4286506b-656c-4767-9814-a9ec579d6403 service nova] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Refreshing network info cache for port 051455f3-4c94-4050-8610-bd8ace9f3988 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 990.441206] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c61aefb1-d0bc-40a0-9a74-b1d68160c9fd tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquiring lock "refresh_cache-d6fef3ef-fec8-4929-b9b8-5e63306aa51d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 990.441206] env[63088]: DEBUG oslo_vmware.api [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285370, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.960706} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.441206] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] c4fb31b2-1301-4a6e-bef2-8ea462caee0d/c4fb31b2-1301-4a6e-bef2-8ea462caee0d.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 990.441206] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: c4fb31b2-1301-4a6e-bef2-8ea462caee0d] Extending root virtual disk to 1048576 {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 990.441206] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a1f10dde-041e-418f-9625-07c354824475 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.441206] env[63088]: DEBUG oslo_vmware.api [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Waiting for the task: (returnval){ [ 990.441206] env[63088]: value = "task-1285371" [ 990.441206] env[63088]: _type = "Task" [ 990.441206] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.441206] env[63088]: DEBUG oslo_vmware.api [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285371, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.464913] env[63088]: DEBUG oslo_concurrency.lockutils [None req-69031b37-7cb7-4292-b542-525ae29cf7c0 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Lock "c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.669s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 990.572660] env[63088]: INFO nova.compute.manager [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3] Rebuilding instance [ 990.575141] env[63088]: DEBUG oslo_concurrency.lockutils [req-4a342f9b-676f-4f0c-a011-bd7d57b2fe6d req-68990000-f6c1-47ba-b896-9c4076b804b5 service nova] Releasing lock "refresh_cache-f582da15-dfc3-45e1-a995-9dd0c9533869" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 990.575437] env[63088]: DEBUG nova.compute.manager [req-4a342f9b-676f-4f0c-a011-bd7d57b2fe6d req-68990000-f6c1-47ba-b896-9c4076b804b5 service nova] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Received event network-changed-b7a01048-9168-46ff-a5b8-83cee3dd3c97 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 990.575694] env[63088]: DEBUG nova.compute.manager [req-4a342f9b-676f-4f0c-a011-bd7d57b2fe6d req-68990000-f6c1-47ba-b896-9c4076b804b5 service nova] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Refreshing instance network info cache due to event network-changed-b7a01048-9168-46ff-a5b8-83cee3dd3c97. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 990.575936] env[63088]: DEBUG oslo_concurrency.lockutils [req-4a342f9b-676f-4f0c-a011-bd7d57b2fe6d req-68990000-f6c1-47ba-b896-9c4076b804b5 service nova] Acquiring lock "refresh_cache-d6fef3ef-fec8-4929-b9b8-5e63306aa51d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 990.614295] env[63088]: DEBUG nova.compute.manager [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 990.615150] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e6ed250-55dc-4176-9f47-48ddf7d6a092 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.672728] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.275s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 990.674936] env[63088]: DEBUG oslo_concurrency.lockutils [None req-82ce7dd7-1f78-4f20-808f-f5b9d77ac35d tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.081s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 990.675202] env[63088]: DEBUG nova.objects.instance [None req-82ce7dd7-1f78-4f20-808f-f5b9d77ac35d tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Lazy-loading 'resources' on Instance uuid 818704d4-2010-4492-8e51-28ad0bd6f570 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 990.701844] env[63088]: INFO nova.network.neutron [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Updating port 79ca8279-a3e2-4792-a4b5-57f899be5c41 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 990.734779] env[63088]: INFO nova.compute.manager [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] Rebuilding instance [ 990.769948] env[63088]: DEBUG nova.compute.manager [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 990.770995] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0933f3aa-096e-4f08-a25e-a7727c7fa25f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.901674] env[63088]: DEBUG oslo_vmware.api [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285371, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063644} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.904150] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: c4fb31b2-1301-4a6e-bef2-8ea462caee0d] Extended root virtual disk {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 990.904942] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0fd76b0-d5fa-4df1-a45e-b4f1eb2c1910 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.928247] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: c4fb31b2-1301-4a6e-bef2-8ea462caee0d] Reconfiguring VM instance instance-00000060 to attach disk [datastore1] c4fb31b2-1301-4a6e-bef2-8ea462caee0d/c4fb31b2-1301-4a6e-bef2-8ea462caee0d.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 990.928695] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-769494ba-d19c-4bc2-9c65-6334339c9567 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.953830] env[63088]: DEBUG oslo_vmware.api [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Waiting for the task: (returnval){ [ 990.953830] env[63088]: value = "task-1285372" [ 990.953830] env[63088]: _type = "Task" [ 990.953830] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.963443] env[63088]: DEBUG oslo_vmware.api [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285372, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.986640] env[63088]: DEBUG nova.network.neutron [req-591139ea-665b-41fa-8be8-e3cbccba845d req-4286506b-656c-4767-9814-a9ec579d6403 service nova] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Added VIF to instance network info cache for port 051455f3-4c94-4050-8610-bd8ace9f3988. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3489}} [ 990.987103] env[63088]: DEBUG nova.network.neutron [req-591139ea-665b-41fa-8be8-e3cbccba845d req-4286506b-656c-4767-9814-a9ec579d6403 service nova] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Updating instance_info_cache with network_info: [{"id": "b7a01048-9168-46ff-a5b8-83cee3dd3c97", "address": "fa:16:3e:e3:57:91", "network": {"id": "1174a284-d3ca-4f9e-aa81-13ee9a693e55", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1994276040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa3d24a1a6c0430985fd80365d986ee1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb7a01048-91", "ovs_interfaceid": "b7a01048-9168-46ff-a5b8-83cee3dd3c97", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "051455f3-4c94-4050-8610-bd8ace9f3988", "address": "fa:16:3e:e5:97:d9", "network": {"id": "1174a284-d3ca-4f9e-aa81-13ee9a693e55", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1994276040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa3d24a1a6c0430985fd80365d986ee1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap051455f3-4c", "ovs_interfaceid": "051455f3-4c94-4050-8610-bd8ace9f3988", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 991.129218] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 991.129540] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1f14f26a-179a-4931-bc51-23dfa201e0fd {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.137190] env[63088]: DEBUG oslo_vmware.api [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Waiting for the task: (returnval){ [ 991.137190] env[63088]: value = "task-1285373" [ 991.137190] env[63088]: _type = "Task" [ 991.137190] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.146114] env[63088]: DEBUG oslo_vmware.api [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Task: {'id': task-1285373, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.283625] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 991.284356] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d0c764e1-c442-41f6-ae9c-496ea823b231 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.294862] env[63088]: DEBUG oslo_vmware.api [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for the task: (returnval){ [ 991.294862] env[63088]: value = "task-1285374" [ 991.294862] env[63088]: _type = "Task" [ 991.294862] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.305070] env[63088]: DEBUG oslo_vmware.api [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285374, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.350949] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e18007f-c447-496d-a2c5-9d19df96f98e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.359530] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87b6b040-b916-4f9b-bd9d-85e6627e2dc5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.393141] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90049212-b5f9-45ca-bf7d-b998bc845bd8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.402564] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05a48361-fb7b-4750-9c57-db3c04ec6f80 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.416712] env[63088]: DEBUG nova.compute.provider_tree [None req-82ce7dd7-1f78-4f20-808f-f5b9d77ac35d tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 991.464430] env[63088]: DEBUG oslo_vmware.api [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285372, 'name': ReconfigVM_Task, 'duration_secs': 0.331681} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.465101] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: c4fb31b2-1301-4a6e-bef2-8ea462caee0d] Reconfigured VM instance instance-00000060 to attach disk [datastore1] c4fb31b2-1301-4a6e-bef2-8ea462caee0d/c4fb31b2-1301-4a6e-bef2-8ea462caee0d.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 991.465379] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-94eac89c-b590-44d3-9521-86ffe819c6dc {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.472534] env[63088]: DEBUG oslo_vmware.api [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Waiting for the task: (returnval){ [ 991.472534] env[63088]: value = "task-1285375" [ 991.472534] env[63088]: _type = "Task" [ 991.472534] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.481323] env[63088]: DEBUG oslo_vmware.api [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285375, 'name': Rename_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.490041] env[63088]: DEBUG oslo_concurrency.lockutils [req-591139ea-665b-41fa-8be8-e3cbccba845d req-4286506b-656c-4767-9814-a9ec579d6403 service nova] Releasing lock "refresh_cache-d6fef3ef-fec8-4929-b9b8-5e63306aa51d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 991.490534] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c61aefb1-d0bc-40a0-9a74-b1d68160c9fd tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquired lock "refresh_cache-d6fef3ef-fec8-4929-b9b8-5e63306aa51d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 991.490730] env[63088]: DEBUG nova.network.neutron [None req-c61aefb1-d0bc-40a0-9a74-b1d68160c9fd tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 991.647564] env[63088]: DEBUG oslo_vmware.api [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Task: {'id': task-1285373, 'name': PowerOffVM_Task, 'duration_secs': 0.156279} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.647844] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 991.648080] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 991.648893] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e912644f-09ab-497c-bc0d-b0d660d3a426 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.655802] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3] Unregistering the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 991.656045] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-14da8edd-187d-497a-8b37-0a6a8c1810c7 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.684423] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3] Unregistered the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 991.684645] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3] Deleting contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 991.684838] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Deleting the datastore file [datastore1] c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3 {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 991.685132] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-785c7bce-2169-4c58-8b47-b1947a5e0f5b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.692775] env[63088]: DEBUG oslo_vmware.api [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Waiting for the task: (returnval){ [ 991.692775] env[63088]: value = "task-1285377" [ 991.692775] env[63088]: _type = "Task" [ 991.692775] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.702603] env[63088]: DEBUG oslo_vmware.api [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Task: {'id': task-1285377, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.805314] env[63088]: DEBUG oslo_vmware.api [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285374, 'name': PowerOffVM_Task, 'duration_secs': 0.191829} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.805614] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 991.805869] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 991.806655] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3cf7074-97f0-483e-919f-e276992077d3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.814356] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] Unregistering the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 991.814589] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bfc5021c-6ba1-44d6-a896-c9f31c7a4cf3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.906237] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] Unregistered the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 991.906476] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] Deleting contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 991.906668] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Deleting the datastore file [datastore1] 69645f50-46fc-4c15-9b39-1721f7636e31 {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 991.906940] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9d8f9093-239a-418b-915d-2b16842b442c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.914890] env[63088]: DEBUG oslo_vmware.api [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for the task: (returnval){ [ 991.914890] env[63088]: value = "task-1285379" [ 991.914890] env[63088]: _type = "Task" [ 991.914890] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.919242] env[63088]: DEBUG nova.scheduler.client.report [None req-82ce7dd7-1f78-4f20-808f-f5b9d77ac35d tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 991.926213] env[63088]: DEBUG oslo_vmware.api [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285379, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.983128] env[63088]: DEBUG oslo_vmware.api [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285375, 'name': Rename_Task, 'duration_secs': 0.154277} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.983481] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: c4fb31b2-1301-4a6e-bef2-8ea462caee0d] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 991.983688] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d0a61d47-1b8e-451a-926d-2511b3a779aa {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.990763] env[63088]: DEBUG oslo_vmware.api [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Waiting for the task: (returnval){ [ 991.990763] env[63088]: value = "task-1285380" [ 991.990763] env[63088]: _type = "Task" [ 991.990763] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.002404] env[63088]: DEBUG oslo_vmware.api [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285380, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.026552] env[63088]: WARNING nova.network.neutron [None req-c61aefb1-d0bc-40a0-9a74-b1d68160c9fd tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] 1174a284-d3ca-4f9e-aa81-13ee9a693e55 already exists in list: networks containing: ['1174a284-d3ca-4f9e-aa81-13ee9a693e55']. ignoring it [ 992.026773] env[63088]: WARNING nova.network.neutron [None req-c61aefb1-d0bc-40a0-9a74-b1d68160c9fd tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] 1174a284-d3ca-4f9e-aa81-13ee9a693e55 already exists in list: networks containing: ['1174a284-d3ca-4f9e-aa81-13ee9a693e55']. ignoring it [ 992.026940] env[63088]: WARNING nova.network.neutron [None req-c61aefb1-d0bc-40a0-9a74-b1d68160c9fd tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] 051455f3-4c94-4050-8610-bd8ace9f3988 already exists in list: port_ids containing: ['051455f3-4c94-4050-8610-bd8ace9f3988']. ignoring it [ 992.185479] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Acquiring lock "refresh_cache-71e37d8e-a454-46c4-a3cc-3d5671a32beb" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 992.185653] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Acquired lock "refresh_cache-71e37d8e-a454-46c4-a3cc-3d5671a32beb" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 992.185894] env[63088]: DEBUG nova.network.neutron [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 992.204049] env[63088]: DEBUG oslo_vmware.api [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Task: {'id': task-1285377, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.370741} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.208025] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 992.208371] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3] Deleted contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 992.208623] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 992.273168] env[63088]: DEBUG nova.compute.manager [req-a8dcd7d6-2f30-44fc-96be-7853a22625e4 req-2e008f80-a78f-49d8-88f6-1073bce97006 service nova] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Received event network-vif-plugged-79ca8279-a3e2-4792-a4b5-57f899be5c41 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 992.273168] env[63088]: DEBUG oslo_concurrency.lockutils [req-a8dcd7d6-2f30-44fc-96be-7853a22625e4 req-2e008f80-a78f-49d8-88f6-1073bce97006 service nova] Acquiring lock "71e37d8e-a454-46c4-a3cc-3d5671a32beb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 992.273168] env[63088]: DEBUG oslo_concurrency.lockutils [req-a8dcd7d6-2f30-44fc-96be-7853a22625e4 req-2e008f80-a78f-49d8-88f6-1073bce97006 service nova] Lock "71e37d8e-a454-46c4-a3cc-3d5671a32beb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 992.273808] env[63088]: DEBUG oslo_concurrency.lockutils [req-a8dcd7d6-2f30-44fc-96be-7853a22625e4 req-2e008f80-a78f-49d8-88f6-1073bce97006 service nova] Lock "71e37d8e-a454-46c4-a3cc-3d5671a32beb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.274132] env[63088]: DEBUG nova.compute.manager [req-a8dcd7d6-2f30-44fc-96be-7853a22625e4 req-2e008f80-a78f-49d8-88f6-1073bce97006 service nova] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] No waiting events found dispatching network-vif-plugged-79ca8279-a3e2-4792-a4b5-57f899be5c41 {{(pid=63088) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 992.274436] env[63088]: WARNING nova.compute.manager [req-a8dcd7d6-2f30-44fc-96be-7853a22625e4 req-2e008f80-a78f-49d8-88f6-1073bce97006 service nova] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Received unexpected event network-vif-plugged-79ca8279-a3e2-4792-a4b5-57f899be5c41 for instance with vm_state shelved_offloaded and task_state spawning. [ 992.274719] env[63088]: DEBUG nova.compute.manager [req-a8dcd7d6-2f30-44fc-96be-7853a22625e4 req-2e008f80-a78f-49d8-88f6-1073bce97006 service nova] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Received event network-changed-79ca8279-a3e2-4792-a4b5-57f899be5c41 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 992.275019] env[63088]: DEBUG nova.compute.manager [req-a8dcd7d6-2f30-44fc-96be-7853a22625e4 req-2e008f80-a78f-49d8-88f6-1073bce97006 service nova] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Refreshing instance network info cache due to event network-changed-79ca8279-a3e2-4792-a4b5-57f899be5c41. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 992.275320] env[63088]: DEBUG oslo_concurrency.lockutils [req-a8dcd7d6-2f30-44fc-96be-7853a22625e4 req-2e008f80-a78f-49d8-88f6-1073bce97006 service nova] Acquiring lock "refresh_cache-71e37d8e-a454-46c4-a3cc-3d5671a32beb" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 992.327606] env[63088]: DEBUG nova.network.neutron [None req-c61aefb1-d0bc-40a0-9a74-b1d68160c9fd tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Updating instance_info_cache with network_info: [{"id": "b7a01048-9168-46ff-a5b8-83cee3dd3c97", "address": "fa:16:3e:e3:57:91", "network": {"id": "1174a284-d3ca-4f9e-aa81-13ee9a693e55", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1994276040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.150", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa3d24a1a6c0430985fd80365d986ee1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb7a01048-91", "ovs_interfaceid": "b7a01048-9168-46ff-a5b8-83cee3dd3c97", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "051455f3-4c94-4050-8610-bd8ace9f3988", "address": "fa:16:3e:e5:97:d9", "network": {"id": "1174a284-d3ca-4f9e-aa81-13ee9a693e55", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1994276040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa3d24a1a6c0430985fd80365d986ee1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap051455f3-4c", "ovs_interfaceid": "051455f3-4c94-4050-8610-bd8ace9f3988", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 992.426054] env[63088]: DEBUG oslo_vmware.api [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285379, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.356445} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.426242] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 992.426400] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] Deleted contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 992.426551] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 992.429698] env[63088]: DEBUG oslo_concurrency.lockutils [None req-82ce7dd7-1f78-4f20-808f-f5b9d77ac35d tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.755s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.432465] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 6.740s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 992.451830] env[63088]: INFO nova.scheduler.client.report [None req-82ce7dd7-1f78-4f20-808f-f5b9d77ac35d tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Deleted allocations for instance 818704d4-2010-4492-8e51-28ad0bd6f570 [ 992.502757] env[63088]: DEBUG oslo_vmware.api [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285380, 'name': PowerOnVM_Task, 'duration_secs': 0.484879} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.503165] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: c4fb31b2-1301-4a6e-bef2-8ea462caee0d] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 992.503208] env[63088]: INFO nova.compute.manager [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: c4fb31b2-1301-4a6e-bef2-8ea462caee0d] Took 7.14 seconds to spawn the instance on the hypervisor. [ 992.503449] env[63088]: DEBUG nova.compute.manager [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: c4fb31b2-1301-4a6e-bef2-8ea462caee0d] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 992.504255] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-beba372a-df4c-4ce3-9444-6cd39bb544aa {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.830974] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c61aefb1-d0bc-40a0-9a74-b1d68160c9fd tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Releasing lock "refresh_cache-d6fef3ef-fec8-4929-b9b8-5e63306aa51d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 992.831615] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c61aefb1-d0bc-40a0-9a74-b1d68160c9fd tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquiring lock "d6fef3ef-fec8-4929-b9b8-5e63306aa51d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 992.831777] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c61aefb1-d0bc-40a0-9a74-b1d68160c9fd tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquired lock "d6fef3ef-fec8-4929-b9b8-5e63306aa51d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 992.832050] env[63088]: DEBUG oslo_concurrency.lockutils [req-4a342f9b-676f-4f0c-a011-bd7d57b2fe6d req-68990000-f6c1-47ba-b896-9c4076b804b5 service nova] Acquired lock "refresh_cache-d6fef3ef-fec8-4929-b9b8-5e63306aa51d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 992.832340] env[63088]: DEBUG nova.network.neutron [req-4a342f9b-676f-4f0c-a011-bd7d57b2fe6d req-68990000-f6c1-47ba-b896-9c4076b804b5 service nova] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Refreshing network info cache for port b7a01048-9168-46ff-a5b8-83cee3dd3c97 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 992.833852] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef780936-c319-46d3-8cba-f7e586643c6c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.852205] env[63088]: DEBUG nova.virt.hardware [None req-c61aefb1-d0bc-40a0-9a74-b1d68160c9fd tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 992.852332] env[63088]: DEBUG nova.virt.hardware [None req-c61aefb1-d0bc-40a0-9a74-b1d68160c9fd tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 992.852496] env[63088]: DEBUG nova.virt.hardware [None req-c61aefb1-d0bc-40a0-9a74-b1d68160c9fd tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 992.852686] env[63088]: DEBUG nova.virt.hardware [None req-c61aefb1-d0bc-40a0-9a74-b1d68160c9fd tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 992.852835] env[63088]: DEBUG nova.virt.hardware [None req-c61aefb1-d0bc-40a0-9a74-b1d68160c9fd tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 992.852982] env[63088]: DEBUG nova.virt.hardware [None req-c61aefb1-d0bc-40a0-9a74-b1d68160c9fd tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 992.853204] env[63088]: DEBUG nova.virt.hardware [None req-c61aefb1-d0bc-40a0-9a74-b1d68160c9fd tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 992.853400] env[63088]: DEBUG nova.virt.hardware [None req-c61aefb1-d0bc-40a0-9a74-b1d68160c9fd tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 992.853525] env[63088]: DEBUG nova.virt.hardware [None req-c61aefb1-d0bc-40a0-9a74-b1d68160c9fd tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 992.853692] env[63088]: DEBUG nova.virt.hardware [None req-c61aefb1-d0bc-40a0-9a74-b1d68160c9fd tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 992.853861] env[63088]: DEBUG nova.virt.hardware [None req-c61aefb1-d0bc-40a0-9a74-b1d68160c9fd tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 992.860095] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-c61aefb1-d0bc-40a0-9a74-b1d68160c9fd tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Reconfiguring VM to attach interface {{(pid=63088) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 992.862935] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cfc2bd7e-435a-43d1-aa8f-1e602995fea7 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.882705] env[63088]: DEBUG oslo_vmware.api [None req-c61aefb1-d0bc-40a0-9a74-b1d68160c9fd tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Waiting for the task: (returnval){ [ 992.882705] env[63088]: value = "task-1285381" [ 992.882705] env[63088]: _type = "Task" [ 992.882705] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.890947] env[63088]: DEBUG oslo_vmware.api [None req-c61aefb1-d0bc-40a0-9a74-b1d68160c9fd tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285381, 'name': ReconfigVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.961936] env[63088]: DEBUG oslo_concurrency.lockutils [None req-82ce7dd7-1f78-4f20-808f-f5b9d77ac35d tempest-ImagesTestJSON-965140468 tempest-ImagesTestJSON-965140468-project-member] Lock "818704d4-2010-4492-8e51-28ad0bd6f570" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.272s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.990730] env[63088]: DEBUG nova.network.neutron [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Updating instance_info_cache with network_info: [{"id": "79ca8279-a3e2-4792-a4b5-57f899be5c41", "address": "fa:16:3e:53:f8:3c", "network": {"id": "cc2c0304-8906-443a-a76e-e7a32158346f", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1867758042-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.139", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b18ddddf2314d83addf550b8cb91977", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0f096917-a0cf-4add-a9d2-23ca1c723b3b", "external-id": "nsx-vlan-transportzone-894", "segmentation_id": 894, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap79ca8279-a3", "ovs_interfaceid": "79ca8279-a3e2-4792-a4b5-57f899be5c41", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 993.022932] env[63088]: INFO nova.compute.manager [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: c4fb31b2-1301-4a6e-bef2-8ea462caee0d] Took 19.84 seconds to build instance. [ 993.243115] env[63088]: DEBUG nova.virt.hardware [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 993.243378] env[63088]: DEBUG nova.virt.hardware [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 993.243541] env[63088]: DEBUG nova.virt.hardware [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 993.243726] env[63088]: DEBUG nova.virt.hardware [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 993.243874] env[63088]: DEBUG nova.virt.hardware [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 993.244035] env[63088]: DEBUG nova.virt.hardware [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 993.244263] env[63088]: DEBUG nova.virt.hardware [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 993.244424] env[63088]: DEBUG nova.virt.hardware [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 993.244599] env[63088]: DEBUG nova.virt.hardware [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 993.244769] env[63088]: DEBUG nova.virt.hardware [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 993.244946] env[63088]: DEBUG nova.virt.hardware [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 993.245841] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-778cda01-fdf1-46d9-a79d-f426ed573337 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.255117] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f32aca0f-79ca-4d8f-8bcc-11d03f068f83 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.274538] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3] Instance VIF info [] {{(pid=63088) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 993.280238] env[63088]: DEBUG oslo.service.loopingcall [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 993.280582] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3] Creating VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 993.281672] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d3d4fdc0-af6d-4d62-9b48-64a9533ec5b7 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.301280] env[63088]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 993.301280] env[63088]: value = "task-1285382" [ 993.301280] env[63088]: _type = "Task" [ 993.301280] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.310703] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285382, 'name': CreateVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.396349] env[63088]: DEBUG oslo_vmware.api [None req-c61aefb1-d0bc-40a0-9a74-b1d68160c9fd tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285381, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.462478] env[63088]: DEBUG nova.virt.hardware [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 993.462743] env[63088]: DEBUG nova.virt.hardware [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 993.462905] env[63088]: DEBUG nova.virt.hardware [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 993.463106] env[63088]: DEBUG nova.virt.hardware [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 993.463259] env[63088]: DEBUG nova.virt.hardware [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 993.463455] env[63088]: DEBUG nova.virt.hardware [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 993.463615] env[63088]: DEBUG nova.virt.hardware [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 993.463779] env[63088]: DEBUG nova.virt.hardware [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 993.463951] env[63088]: DEBUG nova.virt.hardware [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 993.464133] env[63088]: DEBUG nova.virt.hardware [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 993.464315] env[63088]: DEBUG nova.virt.hardware [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 993.465670] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffb94eb9-df60-4760-ba2f-83af2ae975d5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.469274] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance 7b6aadb7-e34b-42b7-b69f-370434f5b665 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 993.469424] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance f582da15-dfc3-45e1-a995-9dd0c9533869 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 993.469546] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance d6fef3ef-fec8-4929-b9b8-5e63306aa51d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 993.469689] env[63088]: WARNING nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance 2ad33254-8030-4454-b023-3359071f85de is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 993.469806] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance 69645f50-46fc-4c15-9b39-1721f7636e31 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 993.469920] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance 71e37d8e-a454-46c4-a3cc-3d5671a32beb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 993.470045] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance 13388466-4adb-4d56-9fc1-e3f5dc516077 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 993.470161] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 993.470272] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance c4fb31b2-1301-4a6e-bef2-8ea462caee0d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 993.479110] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c935b35-acf4-4509-805d-654e95975a7a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.495975] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Releasing lock "refresh_cache-71e37d8e-a454-46c4-a3cc-3d5671a32beb" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 993.498301] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:94:55:ca', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c68b7663-4f0e-47f0-ac7f-40c6d952f7bb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '23b3723c-8ff5-467b-b65b-7eb1ee341733', 'vif_model': 'vmxnet3'}] {{(pid=63088) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 993.506944] env[63088]: DEBUG oslo.service.loopingcall [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 993.509663] env[63088]: DEBUG oslo_concurrency.lockutils [req-a8dcd7d6-2f30-44fc-96be-7853a22625e4 req-2e008f80-a78f-49d8-88f6-1073bce97006 service nova] Acquired lock "refresh_cache-71e37d8e-a454-46c4-a3cc-3d5671a32beb" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 993.509862] env[63088]: DEBUG nova.network.neutron [req-a8dcd7d6-2f30-44fc-96be-7853a22625e4 req-2e008f80-a78f-49d8-88f6-1073bce97006 service nova] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Refreshing network info cache for port 79ca8279-a3e2-4792-a4b5-57f899be5c41 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 993.511373] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] Creating VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 993.512924] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2a73bbe7-22f5-4239-bb2a-b20fd85114c5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.528653] env[63088]: DEBUG oslo_concurrency.lockutils [None req-afb1c06d-61e6-4b7a-ae64-026807ebb932 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lock "c4fb31b2-1301-4a6e-bef2-8ea462caee0d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.355s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.536214] env[63088]: DEBUG nova.virt.hardware [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='79aee9fa8dd788c278273917a4a15fcf',container_format='bare',created_at=2024-10-15T12:20:55Z,direct_url=,disk_format='vmdk',id=e3652b5e-39c5-4fe1-8b63-944e4527acd9,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-34713503-shelved',owner='2b18ddddf2314d83addf550b8cb91977',properties=ImageMetaProps,protected=,size=31668224,status='active',tags=,updated_at=2024-10-15T12:21:10Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 993.536582] env[63088]: DEBUG nova.virt.hardware [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 993.536759] env[63088]: DEBUG nova.virt.hardware [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 993.536950] env[63088]: DEBUG nova.virt.hardware [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 993.537116] env[63088]: DEBUG nova.virt.hardware [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 993.537286] env[63088]: DEBUG nova.virt.hardware [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 993.537506] env[63088]: DEBUG nova.virt.hardware [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 993.537705] env[63088]: DEBUG nova.virt.hardware [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 993.537888] env[63088]: DEBUG nova.virt.hardware [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 993.538070] env[63088]: DEBUG nova.virt.hardware [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 993.538252] env[63088]: DEBUG nova.virt.hardware [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 993.539476] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e898e30a-f8fb-4b58-a31f-ea20e3c1eace {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.545569] env[63088]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 993.545569] env[63088]: value = "task-1285383" [ 993.545569] env[63088]: _type = "Task" [ 993.545569] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.555045] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-981aa6fb-8f79-4a57-b989-f90713ae7dce {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.562264] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285383, 'name': CreateVM_Task} progress is 6%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.575157] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:53:f8:3c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0f096917-a0cf-4add-a9d2-23ca1c723b3b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '79ca8279-a3e2-4792-a4b5-57f899be5c41', 'vif_model': 'vmxnet3'}] {{(pid=63088) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 993.582767] env[63088]: DEBUG oslo.service.loopingcall [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 993.583076] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Creating VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 993.583309] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b3ea6813-1718-4e0c-b962-f92303650a6b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.607623] env[63088]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 993.607623] env[63088]: value = "task-1285384" [ 993.607623] env[63088]: _type = "Task" [ 993.607623] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.616250] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285384, 'name': CreateVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.670131] env[63088]: DEBUG nova.network.neutron [req-4a342f9b-676f-4f0c-a011-bd7d57b2fe6d req-68990000-f6c1-47ba-b896-9c4076b804b5 service nova] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Updated VIF entry in instance network info cache for port b7a01048-9168-46ff-a5b8-83cee3dd3c97. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 993.670643] env[63088]: DEBUG nova.network.neutron [req-4a342f9b-676f-4f0c-a011-bd7d57b2fe6d req-68990000-f6c1-47ba-b896-9c4076b804b5 service nova] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Updating instance_info_cache with network_info: [{"id": "b7a01048-9168-46ff-a5b8-83cee3dd3c97", "address": "fa:16:3e:e3:57:91", "network": {"id": "1174a284-d3ca-4f9e-aa81-13ee9a693e55", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1994276040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.150", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa3d24a1a6c0430985fd80365d986ee1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb7a01048-91", "ovs_interfaceid": "b7a01048-9168-46ff-a5b8-83cee3dd3c97", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "051455f3-4c94-4050-8610-bd8ace9f3988", "address": "fa:16:3e:e5:97:d9", "network": {"id": "1174a284-d3ca-4f9e-aa81-13ee9a693e55", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1994276040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa3d24a1a6c0430985fd80365d986ee1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap051455f3-4c", "ovs_interfaceid": "051455f3-4c94-4050-8610-bd8ace9f3988", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 993.812859] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285382, 'name': CreateVM_Task} progress is 99%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.858800] env[63088]: DEBUG oslo_concurrency.lockutils [None req-818df68b-430f-49fa-83f0-07dc08efcfcd tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquiring lock "c4fb31b2-1301-4a6e-bef2-8ea462caee0d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.859132] env[63088]: DEBUG oslo_concurrency.lockutils [None req-818df68b-430f-49fa-83f0-07dc08efcfcd tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lock "c4fb31b2-1301-4a6e-bef2-8ea462caee0d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.859383] env[63088]: DEBUG oslo_concurrency.lockutils [None req-818df68b-430f-49fa-83f0-07dc08efcfcd tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquiring lock "c4fb31b2-1301-4a6e-bef2-8ea462caee0d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.859665] env[63088]: DEBUG oslo_concurrency.lockutils [None req-818df68b-430f-49fa-83f0-07dc08efcfcd tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lock "c4fb31b2-1301-4a6e-bef2-8ea462caee0d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.859993] env[63088]: DEBUG oslo_concurrency.lockutils [None req-818df68b-430f-49fa-83f0-07dc08efcfcd tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lock "c4fb31b2-1301-4a6e-bef2-8ea462caee0d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.863392] env[63088]: INFO nova.compute.manager [None req-818df68b-430f-49fa-83f0-07dc08efcfcd tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: c4fb31b2-1301-4a6e-bef2-8ea462caee0d] Terminating instance [ 993.865978] env[63088]: DEBUG nova.compute.manager [None req-818df68b-430f-49fa-83f0-07dc08efcfcd tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: c4fb31b2-1301-4a6e-bef2-8ea462caee0d] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 993.866223] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-818df68b-430f-49fa-83f0-07dc08efcfcd tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: c4fb31b2-1301-4a6e-bef2-8ea462caee0d] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 993.867211] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-303c094c-5598-46a5-bca5-5bd38ef812d1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.876837] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-818df68b-430f-49fa-83f0-07dc08efcfcd tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: c4fb31b2-1301-4a6e-bef2-8ea462caee0d] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 993.877185] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3ddc452c-0841-47dd-9c37-315cf291041d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.889018] env[63088]: DEBUG oslo_vmware.api [None req-818df68b-430f-49fa-83f0-07dc08efcfcd tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Waiting for the task: (returnval){ [ 993.889018] env[63088]: value = "task-1285385" [ 993.889018] env[63088]: _type = "Task" [ 993.889018] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.897727] env[63088]: DEBUG oslo_vmware.api [None req-c61aefb1-d0bc-40a0-9a74-b1d68160c9fd tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285381, 'name': ReconfigVM_Task, 'duration_secs': 0.569917} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.898701] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c61aefb1-d0bc-40a0-9a74-b1d68160c9fd tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Releasing lock "d6fef3ef-fec8-4929-b9b8-5e63306aa51d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 993.898937] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-c61aefb1-d0bc-40a0-9a74-b1d68160c9fd tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Reconfigured VM to attach interface {{(pid=63088) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 993.905114] env[63088]: DEBUG oslo_vmware.api [None req-818df68b-430f-49fa-83f0-07dc08efcfcd tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285385, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.974255] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance a39840c1-3d60-478b-987c-2519551963cf has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 994.058134] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285383, 'name': CreateVM_Task} progress is 25%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.117923] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285384, 'name': CreateVM_Task} progress is 25%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.174581] env[63088]: DEBUG oslo_concurrency.lockutils [req-4a342f9b-676f-4f0c-a011-bd7d57b2fe6d req-68990000-f6c1-47ba-b896-9c4076b804b5 service nova] Releasing lock "refresh_cache-d6fef3ef-fec8-4929-b9b8-5e63306aa51d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 994.283611] env[63088]: DEBUG nova.network.neutron [req-a8dcd7d6-2f30-44fc-96be-7853a22625e4 req-2e008f80-a78f-49d8-88f6-1073bce97006 service nova] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Updated VIF entry in instance network info cache for port 79ca8279-a3e2-4792-a4b5-57f899be5c41. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 994.284147] env[63088]: DEBUG nova.network.neutron [req-a8dcd7d6-2f30-44fc-96be-7853a22625e4 req-2e008f80-a78f-49d8-88f6-1073bce97006 service nova] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Updating instance_info_cache with network_info: [{"id": "79ca8279-a3e2-4792-a4b5-57f899be5c41", "address": "fa:16:3e:53:f8:3c", "network": {"id": "cc2c0304-8906-443a-a76e-e7a32158346f", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1867758042-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.139", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b18ddddf2314d83addf550b8cb91977", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0f096917-a0cf-4add-a9d2-23ca1c723b3b", "external-id": "nsx-vlan-transportzone-894", "segmentation_id": 894, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap79ca8279-a3", "ovs_interfaceid": "79ca8279-a3e2-4792-a4b5-57f899be5c41", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 994.312557] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285382, 'name': CreateVM_Task} progress is 99%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.402075] env[63088]: DEBUG oslo_vmware.api [None req-818df68b-430f-49fa-83f0-07dc08efcfcd tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285385, 'name': PowerOffVM_Task, 'duration_secs': 0.201078} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.402478] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-818df68b-430f-49fa-83f0-07dc08efcfcd tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: c4fb31b2-1301-4a6e-bef2-8ea462caee0d] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 994.402735] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-818df68b-430f-49fa-83f0-07dc08efcfcd tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: c4fb31b2-1301-4a6e-bef2-8ea462caee0d] Unregistering the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 994.402946] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2ab79cb2-885e-49f3-9808-62d0a9bcf3ea {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.406935] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c61aefb1-d0bc-40a0-9a74-b1d68160c9fd tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Lock "interface-d6fef3ef-fec8-4929-b9b8-5e63306aa51d-051455f3-4c94-4050-8610-bd8ace9f3988" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.450s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.477692] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance 32418785-8823-4a5e-90a9-e205752530b7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 994.477802] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Total usable vcpus: 48, total allocated vcpus: 8 {{(pid=63088) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 994.477954] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2048MB phys_disk=200GB used_disk=8GB total_vcpus=48 used_vcpus=8 pci_stats=[] {{(pid=63088) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 994.502441] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-818df68b-430f-49fa-83f0-07dc08efcfcd tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: c4fb31b2-1301-4a6e-bef2-8ea462caee0d] Unregistered the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 994.502710] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-818df68b-430f-49fa-83f0-07dc08efcfcd tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: c4fb31b2-1301-4a6e-bef2-8ea462caee0d] Deleting contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 994.503011] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-818df68b-430f-49fa-83f0-07dc08efcfcd tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Deleting the datastore file [datastore1] c4fb31b2-1301-4a6e-bef2-8ea462caee0d {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 994.503226] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5e0abed4-f6ae-4e12-9a32-b795832361f6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.511697] env[63088]: DEBUG oslo_vmware.api [None req-818df68b-430f-49fa-83f0-07dc08efcfcd tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Waiting for the task: (returnval){ [ 994.511697] env[63088]: value = "task-1285387" [ 994.511697] env[63088]: _type = "Task" [ 994.511697] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.523991] env[63088]: DEBUG oslo_vmware.api [None req-818df68b-430f-49fa-83f0-07dc08efcfcd tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285387, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.562179] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285383, 'name': CreateVM_Task} progress is 25%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.620259] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285384, 'name': CreateVM_Task} progress is 25%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.669816] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d6a9284-5d0c-4322-8ee5-d4390ce4deae {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.679920] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b61da289-7d7b-49bb-891e-5bd630e6a32d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.713190] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21fc7b86-c1ce-42ea-9dcf-727e0da60660 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.725193] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ec7fa75-a89d-4c29-8875-04cf99e95154 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.740667] env[63088]: DEBUG nova.compute.provider_tree [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 994.787948] env[63088]: DEBUG oslo_concurrency.lockutils [req-a8dcd7d6-2f30-44fc-96be-7853a22625e4 req-2e008f80-a78f-49d8-88f6-1073bce97006 service nova] Releasing lock "refresh_cache-71e37d8e-a454-46c4-a3cc-3d5671a32beb" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 994.819369] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285382, 'name': CreateVM_Task, 'duration_secs': 1.304268} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.819369] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3] Created VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 994.819921] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 994.820179] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 994.820591] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 994.820848] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bed879cf-486b-40a0-9e67-64453c122e10 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.827201] env[63088]: DEBUG oslo_vmware.api [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Waiting for the task: (returnval){ [ 994.827201] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52967fb0-7b53-14e6-6a87-0193455f58d7" [ 994.827201] env[63088]: _type = "Task" [ 994.827201] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.836322] env[63088]: DEBUG oslo_vmware.api [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52967fb0-7b53-14e6-6a87-0193455f58d7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.025938] env[63088]: DEBUG oslo_vmware.api [None req-818df68b-430f-49fa-83f0-07dc08efcfcd tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285387, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.213976} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.026336] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-818df68b-430f-49fa-83f0-07dc08efcfcd tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 995.026600] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-818df68b-430f-49fa-83f0-07dc08efcfcd tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: c4fb31b2-1301-4a6e-bef2-8ea462caee0d] Deleted contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 995.026877] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-818df68b-430f-49fa-83f0-07dc08efcfcd tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: c4fb31b2-1301-4a6e-bef2-8ea462caee0d] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 995.027158] env[63088]: INFO nova.compute.manager [None req-818df68b-430f-49fa-83f0-07dc08efcfcd tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: c4fb31b2-1301-4a6e-bef2-8ea462caee0d] Took 1.16 seconds to destroy the instance on the hypervisor. [ 995.027502] env[63088]: DEBUG oslo.service.loopingcall [None req-818df68b-430f-49fa-83f0-07dc08efcfcd tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 995.027766] env[63088]: DEBUG nova.compute.manager [-] [instance: c4fb31b2-1301-4a6e-bef2-8ea462caee0d] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 995.027902] env[63088]: DEBUG nova.network.neutron [-] [instance: c4fb31b2-1301-4a6e-bef2-8ea462caee0d] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 995.065657] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285383, 'name': CreateVM_Task, 'duration_secs': 1.322093} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.065894] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] Created VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 995.066748] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 995.119175] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285384, 'name': CreateVM_Task, 'duration_secs': 1.252379} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.119357] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Created VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 995.120050] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e3652b5e-39c5-4fe1-8b63-944e4527acd9" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 995.120197] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e3652b5e-39c5-4fe1-8b63-944e4527acd9" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 995.120590] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e3652b5e-39c5-4fe1-8b63-944e4527acd9" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 995.120852] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6ad78af3-6b59-402f-a472-7e5f3293e8de {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.126548] env[63088]: DEBUG oslo_vmware.api [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Waiting for the task: (returnval){ [ 995.126548] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52901eb3-ce4c-e065-83dd-5e82f36cf2ec" [ 995.126548] env[63088]: _type = "Task" [ 995.126548] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.135189] env[63088]: DEBUG oslo_vmware.api [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52901eb3-ce4c-e065-83dd-5e82f36cf2ec, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.243859] env[63088]: DEBUG nova.scheduler.client.report [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 995.339157] env[63088]: DEBUG oslo_vmware.api [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52967fb0-7b53-14e6-6a87-0193455f58d7, 'name': SearchDatastore_Task, 'duration_secs': 0.010417} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.340294] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 995.340294] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3] Processing image 1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 995.340294] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 995.340294] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 995.340546] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 995.341943] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 995.342472] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 995.342558] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-70b6ca64-9135-4a94-a316-cd843890996c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.345845] env[63088]: DEBUG nova.compute.manager [req-3b824929-4d44-4357-96b7-13eb9e8c8d26 req-35e1e519-a36b-4364-90ba-9902dcf59440 service nova] [instance: c4fb31b2-1301-4a6e-bef2-8ea462caee0d] Received event network-vif-deleted-88bc1e34-07ac-4819-9cd8-d933c1540bb2 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 995.346053] env[63088]: INFO nova.compute.manager [req-3b824929-4d44-4357-96b7-13eb9e8c8d26 req-35e1e519-a36b-4364-90ba-9902dcf59440 service nova] [instance: c4fb31b2-1301-4a6e-bef2-8ea462caee0d] Neutron deleted interface 88bc1e34-07ac-4819-9cd8-d933c1540bb2; detaching it from the instance and deleting it from the info cache [ 995.346234] env[63088]: DEBUG nova.network.neutron [req-3b824929-4d44-4357-96b7-13eb9e8c8d26 req-35e1e519-a36b-4364-90ba-9902dcf59440 service nova] [instance: c4fb31b2-1301-4a6e-bef2-8ea462caee0d] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 995.347225] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-adeaef27-f937-4aca-b692-b43306806007 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.355578] env[63088]: DEBUG oslo_vmware.api [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for the task: (returnval){ [ 995.355578] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52da98a2-b98e-1d67-19f2-039469135415" [ 995.355578] env[63088]: _type = "Task" [ 995.355578] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.360163] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 995.360366] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 995.362014] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-99589e60-8b8f-412d-b255-22748db722fb {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.368197] env[63088]: DEBUG oslo_vmware.api [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52da98a2-b98e-1d67-19f2-039469135415, 'name': SearchDatastore_Task, 'duration_secs': 0.011246} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.369014] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 995.369390] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] Processing image 1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 995.369608] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 995.371936] env[63088]: DEBUG oslo_vmware.api [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Waiting for the task: (returnval){ [ 995.371936] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52338813-fa5d-13de-9e2c-49c26db96219" [ 995.371936] env[63088]: _type = "Task" [ 995.371936] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.380275] env[63088]: DEBUG oslo_vmware.api [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52338813-fa5d-13de-9e2c-49c26db96219, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.637997] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e3652b5e-39c5-4fe1-8b63-944e4527acd9" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 995.638303] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Processing image e3652b5e-39c5-4fe1-8b63-944e4527acd9 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 995.638600] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e3652b5e-39c5-4fe1-8b63-944e4527acd9/e3652b5e-39c5-4fe1-8b63-944e4527acd9.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 995.638773] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e3652b5e-39c5-4fe1-8b63-944e4527acd9/e3652b5e-39c5-4fe1-8b63-944e4527acd9.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 995.638965] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 995.639235] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-63890990-7697-4dcd-a2f4-79024aeb6754 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.649916] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 995.650114] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 995.650862] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9440efa6-f39f-4e27-af13-9fbeb513ffa6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.656429] env[63088]: DEBUG oslo_vmware.api [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Waiting for the task: (returnval){ [ 995.656429] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]525e6e6c-b5e3-25af-f718-7e89f32c29fc" [ 995.656429] env[63088]: _type = "Task" [ 995.656429] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.665190] env[63088]: DEBUG oslo_vmware.api [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]525e6e6c-b5e3-25af-f718-7e89f32c29fc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.738989] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4faf3ac2-8861-4d6a-8416-593c2f5689cf tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquiring lock "interface-d6fef3ef-fec8-4929-b9b8-5e63306aa51d-051455f3-4c94-4050-8610-bd8ace9f3988" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 995.739319] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4faf3ac2-8861-4d6a-8416-593c2f5689cf tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Lock "interface-d6fef3ef-fec8-4929-b9b8-5e63306aa51d-051455f3-4c94-4050-8610-bd8ace9f3988" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 995.752187] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63088) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 995.752411] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.320s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.753110] env[63088]: DEBUG oslo_concurrency.lockutils [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.103s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 995.754619] env[63088]: INFO nova.compute.claims [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: a39840c1-3d60-478b-987c-2519551963cf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 995.819013] env[63088]: DEBUG nova.network.neutron [-] [instance: c4fb31b2-1301-4a6e-bef2-8ea462caee0d] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 995.850511] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d851d763-8d28-4213-be78-4b1731a58029 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.861110] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7caeb2d6-2ae1-43f4-a698-97c19165a3b8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.881544] env[63088]: DEBUG oslo_vmware.api [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52338813-fa5d-13de-9e2c-49c26db96219, 'name': SearchDatastore_Task, 'duration_secs': 0.009323} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.889097] env[63088]: DEBUG nova.compute.manager [req-3b824929-4d44-4357-96b7-13eb9e8c8d26 req-35e1e519-a36b-4364-90ba-9902dcf59440 service nova] [instance: c4fb31b2-1301-4a6e-bef2-8ea462caee0d] Detach interface failed, port_id=88bc1e34-07ac-4819-9cd8-d933c1540bb2, reason: Instance c4fb31b2-1301-4a6e-bef2-8ea462caee0d could not be found. {{(pid=63088) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 995.889498] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a2cb4487-939a-4e67-b3de-007ca3a28ceb {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.895529] env[63088]: DEBUG oslo_vmware.api [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Waiting for the task: (returnval){ [ 995.895529] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]5273fe13-493a-b117-3872-ffbcb2c5803b" [ 995.895529] env[63088]: _type = "Task" [ 995.895529] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.904688] env[63088]: DEBUG oslo_vmware.api [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]5273fe13-493a-b117-3872-ffbcb2c5803b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.167372] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Preparing fetch location {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 996.167635] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Fetch image to [datastore2] OSTACK_IMG_12153d93-69bb-4781-9587-7b1345b46796/OSTACK_IMG_12153d93-69bb-4781-9587-7b1345b46796.vmdk {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 996.167824] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Downloading stream optimized image e3652b5e-39c5-4fe1-8b63-944e4527acd9 to [datastore2] OSTACK_IMG_12153d93-69bb-4781-9587-7b1345b46796/OSTACK_IMG_12153d93-69bb-4781-9587-7b1345b46796.vmdk on the data store datastore2 as vApp {{(pid=63088) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 996.168063] env[63088]: DEBUG nova.virt.vmwareapi.images [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Downloading image file data e3652b5e-39c5-4fe1-8b63-944e4527acd9 to the ESX as VM named 'OSTACK_IMG_12153d93-69bb-4781-9587-7b1345b46796' {{(pid=63088) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 996.245910] env[63088]: DEBUG oslo_vmware.rw_handles [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 996.245910] env[63088]: value = "resgroup-9" [ 996.245910] env[63088]: _type = "ResourcePool" [ 996.245910] env[63088]: }. {{(pid=63088) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 996.245910] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4faf3ac2-8861-4d6a-8416-593c2f5689cf tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquiring lock "d6fef3ef-fec8-4929-b9b8-5e63306aa51d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 996.245910] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4faf3ac2-8861-4d6a-8416-593c2f5689cf tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquired lock "d6fef3ef-fec8-4929-b9b8-5e63306aa51d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 996.245910] env[63088]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-f779f64d-58ac-4785-812a-854ce3845426 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.261255] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d41df28-9e9a-41f1-9ca6-8ff74c726ee9 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.286634] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89ab2ca3-ac61-471c-b975-ef2a6a3a1fb8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.289704] env[63088]: DEBUG oslo_vmware.rw_handles [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Lease: (returnval){ [ 996.289704] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52d7487d-d1c8-dc65-04c1-2b21e0f1a556" [ 996.289704] env[63088]: _type = "HttpNfcLease" [ 996.289704] env[63088]: } obtained for vApp import into resource pool (val){ [ 996.289704] env[63088]: value = "resgroup-9" [ 996.289704] env[63088]: _type = "ResourcePool" [ 996.289704] env[63088]: }. {{(pid=63088) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 996.290361] env[63088]: DEBUG oslo_vmware.api [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Waiting for the lease: (returnval){ [ 996.290361] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52d7487d-d1c8-dc65-04c1-2b21e0f1a556" [ 996.290361] env[63088]: _type = "HttpNfcLease" [ 996.290361] env[63088]: } to be ready. {{(pid=63088) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 997.033842] env[63088]: INFO nova.compute.manager [-] [instance: c4fb31b2-1301-4a6e-bef2-8ea462caee0d] Took 2.01 seconds to deallocate network for instance. [ 997.042727] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-4faf3ac2-8861-4d6a-8416-593c2f5689cf tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Reconfiguring VM to detach interface {{(pid=63088) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1974}} [ 997.048864] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-716b5df1-d8e7-4d9e-9e24-4691b7b2dd7b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.073232] env[63088]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 997.073232] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52d7487d-d1c8-dc65-04c1-2b21e0f1a556" [ 997.073232] env[63088]: _type = "HttpNfcLease" [ 997.073232] env[63088]: } is initializing. {{(pid=63088) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 997.073232] env[63088]: WARNING oslo_vmware.common.loopingcall [-] task run outlasted interval by 0.27879200000000004 sec [ 997.081666] env[63088]: DEBUG oslo_vmware.api [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]5273fe13-493a-b117-3872-ffbcb2c5803b, 'name': SearchDatastore_Task, 'duration_secs': 0.009853} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.081666] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 997.081666] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore2] c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3/c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 997.081666] env[63088]: DEBUG oslo_vmware.api [None req-4faf3ac2-8861-4d6a-8416-593c2f5689cf tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Waiting for the task: (returnval){ [ 997.081666] env[63088]: value = "task-1285389" [ 997.081666] env[63088]: _type = "Task" [ 997.081666] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.081666] env[63088]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 997.081666] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52d7487d-d1c8-dc65-04c1-2b21e0f1a556" [ 997.081666] env[63088]: _type = "HttpNfcLease" [ 997.081666] env[63088]: } is initializing. {{(pid=63088) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 997.082202] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 997.082202] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 997.082325] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-559f4446-0f44-432b-a0f8-e3ba802d8035 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.084400] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2460ff64-f5cf-4545-89e9-08cc867522fa {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.100286] env[63088]: DEBUG oslo_vmware.api [None req-4faf3ac2-8861-4d6a-8416-593c2f5689cf tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285389, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.100286] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 997.100286] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 997.100942] env[63088]: DEBUG oslo_vmware.api [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Waiting for the task: (returnval){ [ 997.100942] env[63088]: value = "task-1285390" [ 997.100942] env[63088]: _type = "Task" [ 997.100942] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.101202] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-61383fc1-3a48-4d8c-a547-5ecdcf5a29cb {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.118300] env[63088]: DEBUG oslo_vmware.api [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Task: {'id': task-1285390, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.118716] env[63088]: DEBUG oslo_vmware.api [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for the task: (returnval){ [ 997.118716] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52942af5-1c8b-4a83-bf1e-032c979aef20" [ 997.118716] env[63088]: _type = "Task" [ 997.118716] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.129971] env[63088]: DEBUG oslo_vmware.api [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52942af5-1c8b-4a83-bf1e-032c979aef20, 'name': SearchDatastore_Task, 'duration_secs': 0.009596} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.133673] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bdcd6295-48aa-4b83-af47-c182f9d05283 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.139591] env[63088]: DEBUG oslo_vmware.api [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for the task: (returnval){ [ 997.139591] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]5256e3ae-0cf8-066c-6009-bd5d9340878c" [ 997.139591] env[63088]: _type = "Task" [ 997.139591] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.152569] env[63088]: DEBUG oslo_vmware.api [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]5256e3ae-0cf8-066c-6009-bd5d9340878c, 'name': SearchDatastore_Task, 'duration_secs': 0.010518} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.152852] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 997.153134] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore2] 69645f50-46fc-4c15-9b39-1721f7636e31/69645f50-46fc-4c15-9b39-1721f7636e31.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 997.153407] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3a53189f-099d-4171-b6fa-17ac6bde1045 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.164089] env[63088]: DEBUG oslo_vmware.api [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for the task: (returnval){ [ 997.164089] env[63088]: value = "task-1285391" [ 997.164089] env[63088]: _type = "Task" [ 997.164089] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.173289] env[63088]: DEBUG oslo_vmware.api [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285391, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.237137] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7dc9ddc-0857-4820-915d-b7a0863ca1cd {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.246878] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b79b0416-38ef-4d4e-93b3-58ce7da12162 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.281873] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53359b91-3826-4a46-b2ba-df610233ec99 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.290843] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1e80cbd-cc4f-49ee-907a-08064fa1757d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.307200] env[63088]: DEBUG nova.compute.provider_tree [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 997.572083] env[63088]: DEBUG oslo_concurrency.lockutils [None req-818df68b-430f-49fa-83f0-07dc08efcfcd tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 997.579803] env[63088]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 997.579803] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52d7487d-d1c8-dc65-04c1-2b21e0f1a556" [ 997.579803] env[63088]: _type = "HttpNfcLease" [ 997.579803] env[63088]: } is ready. {{(pid=63088) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 997.580213] env[63088]: DEBUG oslo_vmware.rw_handles [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 997.580213] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52d7487d-d1c8-dc65-04c1-2b21e0f1a556" [ 997.580213] env[63088]: _type = "HttpNfcLease" [ 997.580213] env[63088]: }. {{(pid=63088) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 997.581098] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7ca3e6c-7f40-4a7c-8edb-6c21d3f73d2c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.595075] env[63088]: DEBUG oslo_vmware.rw_handles [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52a177ec-9ca1-db85-fd1b-685cf648c8cd/disk-0.vmdk from lease info. {{(pid=63088) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 997.595385] env[63088]: DEBUG oslo_vmware.rw_handles [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Creating HTTP connection to write to file with size = 31668224 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52a177ec-9ca1-db85-fd1b-685cf648c8cd/disk-0.vmdk. {{(pid=63088) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 997.600797] env[63088]: DEBUG oslo_vmware.api [None req-4faf3ac2-8861-4d6a-8416-593c2f5689cf tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285389, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.670647] env[63088]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-9d6cc5fc-c500-4d3c-bc8c-3a9bea51bd5d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.675359] env[63088]: DEBUG oslo_vmware.api [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Task: {'id': task-1285390, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.472309} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.675649] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore2] c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3/c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 997.675864] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3] Extending root virtual disk to 1048576 {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 997.676804] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-68a264f1-838a-4014-bf63-ae4cffbd393c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.683356] env[63088]: DEBUG oslo_vmware.api [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285391, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.685810] env[63088]: DEBUG oslo_vmware.api [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Waiting for the task: (returnval){ [ 997.685810] env[63088]: value = "task-1285392" [ 997.685810] env[63088]: _type = "Task" [ 997.685810] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.696191] env[63088]: DEBUG oslo_vmware.api [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Task: {'id': task-1285392, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.811052] env[63088]: DEBUG nova.scheduler.client.report [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 998.100785] env[63088]: DEBUG oslo_vmware.api [None req-4faf3ac2-8861-4d6a-8416-593c2f5689cf tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285389, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.177701] env[63088]: DEBUG oslo_vmware.api [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285391, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.704864} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.178137] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore2] 69645f50-46fc-4c15-9b39-1721f7636e31/69645f50-46fc-4c15-9b39-1721f7636e31.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 998.178457] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] Extending root virtual disk to 1048576 {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 998.181503] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-804cb3b1-83d4-4424-967c-f60efa4cc359 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.198073] env[63088]: DEBUG oslo_vmware.api [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for the task: (returnval){ [ 998.198073] env[63088]: value = "task-1285393" [ 998.198073] env[63088]: _type = "Task" [ 998.198073] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.208482] env[63088]: DEBUG oslo_vmware.api [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Task: {'id': task-1285392, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.17967} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.217593] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3] Extended root virtual disk {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 998.217593] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52beccd0-9f27-4bc0-b4dd-06e292bbbaf2 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.228444] env[63088]: DEBUG oslo_vmware.api [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285393, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.248505] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3] Reconfiguring VM instance instance-0000005f to attach disk [datastore2] c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3/c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 998.253020] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-206bb652-77d7-4d20-96f1-12c078ee92d1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.277462] env[63088]: DEBUG oslo_vmware.api [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Waiting for the task: (returnval){ [ 998.277462] env[63088]: value = "task-1285394" [ 998.277462] env[63088]: _type = "Task" [ 998.277462] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.290302] env[63088]: DEBUG oslo_vmware.api [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Task: {'id': task-1285394, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.318224] env[63088]: DEBUG oslo_concurrency.lockutils [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.564s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 998.318224] env[63088]: DEBUG nova.compute.manager [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: a39840c1-3d60-478b-987c-2519551963cf] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 998.321975] env[63088]: DEBUG oslo_concurrency.lockutils [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.803s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 998.323249] env[63088]: INFO nova.compute.claims [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 32418785-8823-4a5e-90a9-e205752530b7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 998.597749] env[63088]: DEBUG oslo_vmware.api [None req-4faf3ac2-8861-4d6a-8416-593c2f5689cf tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285389, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.709453] env[63088]: DEBUG oslo_vmware.api [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285393, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.085879} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.709743] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] Extended root virtual disk {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 998.710603] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d01063cd-38bb-4982-8fc7-de7473b790a2 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.735595] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] Reconfiguring VM instance instance-0000005d to attach disk [datastore2] 69645f50-46fc-4c15-9b39-1721f7636e31/69645f50-46fc-4c15-9b39-1721f7636e31.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 998.741334] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-645690e9-8cae-478a-9f1d-b5a3db47f8fe {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.766208] env[63088]: DEBUG oslo_vmware.api [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for the task: (returnval){ [ 998.766208] env[63088]: value = "task-1285395" [ 998.766208] env[63088]: _type = "Task" [ 998.766208] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.776278] env[63088]: DEBUG oslo_vmware.api [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285395, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.789290] env[63088]: DEBUG oslo_vmware.api [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Task: {'id': task-1285394, 'name': ReconfigVM_Task, 'duration_secs': 0.430474} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.789606] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3] Reconfigured VM instance instance-0000005f to attach disk [datastore2] c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3/c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 998.790238] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-778facff-28d9-4d16-b89d-c11c94f7e2fe {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.799261] env[63088]: DEBUG oslo_vmware.api [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Waiting for the task: (returnval){ [ 998.799261] env[63088]: value = "task-1285396" [ 998.799261] env[63088]: _type = "Task" [ 998.799261] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.811134] env[63088]: DEBUG oslo_vmware.api [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Task: {'id': task-1285396, 'name': Rename_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.827921] env[63088]: DEBUG nova.compute.utils [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 998.829513] env[63088]: DEBUG nova.compute.manager [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: a39840c1-3d60-478b-987c-2519551963cf] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 998.829707] env[63088]: DEBUG nova.network.neutron [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: a39840c1-3d60-478b-987c-2519551963cf] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 998.881053] env[63088]: DEBUG nova.policy [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4424adc5bbe34a91b669c3a907882644', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f33f2701fad94864a8c406a404bc0a42', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 998.909678] env[63088]: DEBUG oslo_vmware.rw_handles [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Completed reading data from the image iterator. {{(pid=63088) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 998.909979] env[63088]: DEBUG oslo_vmware.rw_handles [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52a177ec-9ca1-db85-fd1b-685cf648c8cd/disk-0.vmdk. {{(pid=63088) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 998.910912] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89cb4cd0-f96d-4070-a278-1560b71c776e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.918455] env[63088]: DEBUG oslo_vmware.rw_handles [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52a177ec-9ca1-db85-fd1b-685cf648c8cd/disk-0.vmdk is in state: ready. {{(pid=63088) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 998.918648] env[63088]: DEBUG oslo_vmware.rw_handles [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52a177ec-9ca1-db85-fd1b-685cf648c8cd/disk-0.vmdk. {{(pid=63088) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 998.918888] env[63088]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-10170400-25bc-45bd-b55e-bf1286e035d8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.098720] env[63088]: DEBUG oslo_vmware.api [None req-4faf3ac2-8861-4d6a-8416-593c2f5689cf tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285389, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.176377] env[63088]: DEBUG oslo_vmware.rw_handles [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52a177ec-9ca1-db85-fd1b-685cf648c8cd/disk-0.vmdk. {{(pid=63088) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 999.176675] env[63088]: INFO nova.virt.vmwareapi.images [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Downloaded image file data e3652b5e-39c5-4fe1-8b63-944e4527acd9 [ 999.177724] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87faac69-7338-490a-a227-a8bb74dde3b5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.195916] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1057676f-9dfd-4976-86de-7ac9c14a47a7 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.200671] env[63088]: DEBUG nova.network.neutron [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: a39840c1-3d60-478b-987c-2519551963cf] Successfully created port: f19d02ac-6732-44b5-b454-df5777588a7e {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 999.224454] env[63088]: INFO nova.virt.vmwareapi.images [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] The imported VM was unregistered [ 999.227296] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Caching image {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 999.227538] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Creating directory with path [datastore2] devstack-image-cache_base/e3652b5e-39c5-4fe1-8b63-944e4527acd9 {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 999.227998] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9ceac392-f7cf-46cf-8ca0-dd321dc1b0d4 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.238843] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Created directory with path [datastore2] devstack-image-cache_base/e3652b5e-39c5-4fe1-8b63-944e4527acd9 {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 999.239067] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_12153d93-69bb-4781-9587-7b1345b46796/OSTACK_IMG_12153d93-69bb-4781-9587-7b1345b46796.vmdk to [datastore2] devstack-image-cache_base/e3652b5e-39c5-4fe1-8b63-944e4527acd9/e3652b5e-39c5-4fe1-8b63-944e4527acd9.vmdk. {{(pid=63088) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 999.239344] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-020ee5e1-4aae-4c7a-afc8-8cc1bfc675a6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.247995] env[63088]: DEBUG oslo_vmware.api [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Waiting for the task: (returnval){ [ 999.247995] env[63088]: value = "task-1285398" [ 999.247995] env[63088]: _type = "Task" [ 999.247995] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.270318] env[63088]: DEBUG oslo_vmware.api [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285398, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.279798] env[63088]: DEBUG oslo_vmware.api [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285395, 'name': ReconfigVM_Task, 'duration_secs': 0.420603} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.280099] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] Reconfigured VM instance instance-0000005d to attach disk [datastore2] 69645f50-46fc-4c15-9b39-1721f7636e31/69645f50-46fc-4c15-9b39-1721f7636e31.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 999.281167] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4338c472-3eee-40c6-acf4-9415f76330ab {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.289050] env[63088]: DEBUG oslo_vmware.api [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for the task: (returnval){ [ 999.289050] env[63088]: value = "task-1285399" [ 999.289050] env[63088]: _type = "Task" [ 999.289050] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.299265] env[63088]: DEBUG oslo_vmware.api [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285399, 'name': Rename_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.311241] env[63088]: DEBUG oslo_vmware.api [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Task: {'id': task-1285396, 'name': Rename_Task, 'duration_secs': 0.19225} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.311799] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 999.312835] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-95e2dedd-1258-42d9-889a-658939642858 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.319360] env[63088]: DEBUG oslo_vmware.api [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Waiting for the task: (returnval){ [ 999.319360] env[63088]: value = "task-1285400" [ 999.319360] env[63088]: _type = "Task" [ 999.319360] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.327938] env[63088]: DEBUG oslo_vmware.api [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Task: {'id': task-1285400, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.334639] env[63088]: DEBUG nova.compute.manager [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: a39840c1-3d60-478b-987c-2519551963cf] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 999.531850] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89822503-0fbd-41c8-8971-a696a4f9f398 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.541233] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a31bf10-06ed-4dc3-979a-c5dcbab73135 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.549614] env[63088]: DEBUG oslo_concurrency.lockutils [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquiring lock "2b87a9f1-5f10-43c2-8bc9-6d560ec88015" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.549886] env[63088]: DEBUG oslo_concurrency.lockutils [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lock "2b87a9f1-5f10-43c2-8bc9-6d560ec88015" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.581111] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-043bddf2-bd61-4635-acf1-8e08b7270484 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.593694] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7f4963a-3267-4cf6-92e9-e7f9067d8f75 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.604301] env[63088]: DEBUG oslo_vmware.api [None req-4faf3ac2-8861-4d6a-8416-593c2f5689cf tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285389, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.613816] env[63088]: DEBUG nova.compute.provider_tree [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 999.761821] env[63088]: DEBUG oslo_vmware.api [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285398, 'name': MoveVirtualDisk_Task} progress is 18%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.800182] env[63088]: DEBUG oslo_vmware.api [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285399, 'name': Rename_Task, 'duration_secs': 0.221793} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.800592] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 999.800878] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ff31fbdf-6dd8-43ae-b9a6-97762b607316 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.808472] env[63088]: DEBUG oslo_vmware.api [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for the task: (returnval){ [ 999.808472] env[63088]: value = "task-1285401" [ 999.808472] env[63088]: _type = "Task" [ 999.808472] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.817875] env[63088]: DEBUG oslo_vmware.api [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285401, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.829985] env[63088]: DEBUG oslo_vmware.api [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Task: {'id': task-1285400, 'name': PowerOnVM_Task} progress is 71%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.054077] env[63088]: DEBUG nova.compute.manager [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1000.101101] env[63088]: DEBUG oslo_vmware.api [None req-4faf3ac2-8861-4d6a-8416-593c2f5689cf tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285389, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.116976] env[63088]: DEBUG nova.scheduler.client.report [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1000.262016] env[63088]: DEBUG oslo_vmware.api [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285398, 'name': MoveVirtualDisk_Task} progress is 40%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.321036] env[63088]: DEBUG oslo_vmware.api [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285401, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.330768] env[63088]: DEBUG oslo_vmware.api [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Task: {'id': task-1285400, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.346286] env[63088]: DEBUG nova.compute.manager [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: a39840c1-3d60-478b-987c-2519551963cf] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1000.375174] env[63088]: DEBUG nova.virt.hardware [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1000.375507] env[63088]: DEBUG nova.virt.hardware [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1000.375674] env[63088]: DEBUG nova.virt.hardware [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1000.375890] env[63088]: DEBUG nova.virt.hardware [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1000.376217] env[63088]: DEBUG nova.virt.hardware [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1000.376467] env[63088]: DEBUG nova.virt.hardware [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1000.376699] env[63088]: DEBUG nova.virt.hardware [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1000.376869] env[63088]: DEBUG nova.virt.hardware [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1000.377060] env[63088]: DEBUG nova.virt.hardware [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1000.377235] env[63088]: DEBUG nova.virt.hardware [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1000.377455] env[63088]: DEBUG nova.virt.hardware [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1000.378530] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4999b453-4c01-44a5-bc76-7b8c24d196d1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.388139] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16021424-cc4d-4f73-b233-70e15d3a6e8e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.580389] env[63088]: DEBUG oslo_concurrency.lockutils [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.601785] env[63088]: DEBUG oslo_vmware.api [None req-4faf3ac2-8861-4d6a-8416-593c2f5689cf tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285389, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.622574] env[63088]: DEBUG oslo_concurrency.lockutils [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.301s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.623236] env[63088]: DEBUG nova.compute.manager [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 32418785-8823-4a5e-90a9-e205752530b7] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1000.626851] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0de4c45b-7355-4582-b6ec-c78d8eef867d tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.573s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.627102] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0de4c45b-7355-4582-b6ec-c78d8eef867d tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.629381] env[63088]: DEBUG oslo_concurrency.lockutils [None req-818df68b-430f-49fa-83f0-07dc08efcfcd tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.057s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.629665] env[63088]: DEBUG nova.objects.instance [None req-818df68b-430f-49fa-83f0-07dc08efcfcd tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lazy-loading 'resources' on Instance uuid c4fb31b2-1301-4a6e-bef2-8ea462caee0d {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1000.651557] env[63088]: INFO nova.scheduler.client.report [None req-0de4c45b-7355-4582-b6ec-c78d8eef867d tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Deleted allocations for instance 2ad33254-8030-4454-b023-3359071f85de [ 1000.760670] env[63088]: DEBUG oslo_vmware.api [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285398, 'name': MoveVirtualDisk_Task} progress is 63%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.819980] env[63088]: DEBUG oslo_vmware.api [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285401, 'name': PowerOnVM_Task} progress is 71%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.830939] env[63088]: DEBUG oslo_vmware.api [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Task: {'id': task-1285400, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.924050] env[63088]: DEBUG nova.compute.manager [req-9e5395b7-db69-4192-b6d5-d5540db27ace req-275fdc79-cc46-4794-bb64-83757019d2de service nova] [instance: a39840c1-3d60-478b-987c-2519551963cf] Received event network-vif-plugged-f19d02ac-6732-44b5-b454-df5777588a7e {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1000.924351] env[63088]: DEBUG oslo_concurrency.lockutils [req-9e5395b7-db69-4192-b6d5-d5540db27ace req-275fdc79-cc46-4794-bb64-83757019d2de service nova] Acquiring lock "a39840c1-3d60-478b-987c-2519551963cf-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.924691] env[63088]: DEBUG oslo_concurrency.lockutils [req-9e5395b7-db69-4192-b6d5-d5540db27ace req-275fdc79-cc46-4794-bb64-83757019d2de service nova] Lock "a39840c1-3d60-478b-987c-2519551963cf-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.924740] env[63088]: DEBUG oslo_concurrency.lockutils [req-9e5395b7-db69-4192-b6d5-d5540db27ace req-275fdc79-cc46-4794-bb64-83757019d2de service nova] Lock "a39840c1-3d60-478b-987c-2519551963cf-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.924906] env[63088]: DEBUG nova.compute.manager [req-9e5395b7-db69-4192-b6d5-d5540db27ace req-275fdc79-cc46-4794-bb64-83757019d2de service nova] [instance: a39840c1-3d60-478b-987c-2519551963cf] No waiting events found dispatching network-vif-plugged-f19d02ac-6732-44b5-b454-df5777588a7e {{(pid=63088) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1000.925092] env[63088]: WARNING nova.compute.manager [req-9e5395b7-db69-4192-b6d5-d5540db27ace req-275fdc79-cc46-4794-bb64-83757019d2de service nova] [instance: a39840c1-3d60-478b-987c-2519551963cf] Received unexpected event network-vif-plugged-f19d02ac-6732-44b5-b454-df5777588a7e for instance with vm_state building and task_state spawning. [ 1001.076398] env[63088]: DEBUG nova.network.neutron [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: a39840c1-3d60-478b-987c-2519551963cf] Successfully updated port: f19d02ac-6732-44b5-b454-df5777588a7e {{(pid=63088) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1001.106935] env[63088]: DEBUG oslo_vmware.api [None req-4faf3ac2-8861-4d6a-8416-593c2f5689cf tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285389, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.129460] env[63088]: DEBUG nova.compute.utils [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1001.129949] env[63088]: DEBUG nova.compute.manager [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 32418785-8823-4a5e-90a9-e205752530b7] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1001.130225] env[63088]: DEBUG nova.network.neutron [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 32418785-8823-4a5e-90a9-e205752530b7] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1001.160654] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0de4c45b-7355-4582-b6ec-c78d8eef867d tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Lock "2ad33254-8030-4454-b023-3359071f85de" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.014s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.187312] env[63088]: DEBUG nova.policy [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '139b65d9cf314b3db1d685d63bc19bba', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e48f62dee9ad4e9b94b67bd871db5c63', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 1001.276148] env[63088]: DEBUG oslo_vmware.api [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285398, 'name': MoveVirtualDisk_Task} progress is 83%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.322602] env[63088]: DEBUG oslo_vmware.api [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285401, 'name': PowerOnVM_Task} progress is 78%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.325490] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ad2a493-5886-490e-b051-8dfee06d1cef {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.338840] env[63088]: DEBUG oslo_vmware.api [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Task: {'id': task-1285400, 'name': PowerOnVM_Task, 'duration_secs': 1.962177} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.340740] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1001.341046] env[63088]: DEBUG nova.compute.manager [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1001.342011] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6093f0b8-35a9-4f96-bff0-223c8f20d7eb {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.345620] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-657defd6-74f5-44d6-b47a-88b133bf1dcb {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.380312] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0ecd594-e121-45df-a643-fd02851cc9d9 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.392600] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f0d1ce5-2eda-4d14-b328-294451812fc2 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.408064] env[63088]: DEBUG nova.compute.provider_tree [None req-818df68b-430f-49fa-83f0-07dc08efcfcd tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1001.509684] env[63088]: DEBUG nova.network.neutron [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 32418785-8823-4a5e-90a9-e205752530b7] Successfully created port: 1e6d9172-35f9-4d1f-97ee-409ada1e1cd0 {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1001.579094] env[63088]: DEBUG oslo_concurrency.lockutils [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquiring lock "refresh_cache-a39840c1-3d60-478b-987c-2519551963cf" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1001.579311] env[63088]: DEBUG oslo_concurrency.lockutils [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquired lock "refresh_cache-a39840c1-3d60-478b-987c-2519551963cf" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1001.579508] env[63088]: DEBUG nova.network.neutron [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: a39840c1-3d60-478b-987c-2519551963cf] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1001.603624] env[63088]: DEBUG oslo_vmware.api [None req-4faf3ac2-8861-4d6a-8416-593c2f5689cf tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285389, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.633371] env[63088]: DEBUG nova.compute.manager [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 32418785-8823-4a5e-90a9-e205752530b7] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1001.764973] env[63088]: DEBUG oslo_vmware.api [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285398, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.821548] env[63088]: DEBUG oslo_vmware.api [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285401, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.894273] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1001.911477] env[63088]: DEBUG nova.scheduler.client.report [None req-818df68b-430f-49fa-83f0-07dc08efcfcd tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1002.102639] env[63088]: DEBUG oslo_vmware.api [None req-4faf3ac2-8861-4d6a-8416-593c2f5689cf tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285389, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.112290] env[63088]: DEBUG nova.network.neutron [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: a39840c1-3d60-478b-987c-2519551963cf] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1002.263921] env[63088]: DEBUG oslo_vmware.api [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285398, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.677439} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.264252] env[63088]: INFO nova.virt.vmwareapi.ds_util [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_12153d93-69bb-4781-9587-7b1345b46796/OSTACK_IMG_12153d93-69bb-4781-9587-7b1345b46796.vmdk to [datastore2] devstack-image-cache_base/e3652b5e-39c5-4fe1-8b63-944e4527acd9/e3652b5e-39c5-4fe1-8b63-944e4527acd9.vmdk. [ 1002.264456] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Cleaning up location [datastore2] OSTACK_IMG_12153d93-69bb-4781-9587-7b1345b46796 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1002.264625] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_12153d93-69bb-4781-9587-7b1345b46796 {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1002.264884] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b3aa34f2-7cfa-4b7b-863b-207e57a8e939 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.271828] env[63088]: DEBUG nova.network.neutron [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: a39840c1-3d60-478b-987c-2519551963cf] Updating instance_info_cache with network_info: [{"id": "f19d02ac-6732-44b5-b454-df5777588a7e", "address": "fa:16:3e:14:a8:cd", "network": {"id": "dff14a7f-0af1-4e4e-a498-86d7c9816e6b", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-762073671-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f33f2701fad94864a8c406a404bc0a42", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "21310d90-efbc-45a8-a97f-c4358606530f", "external-id": "nsx-vlan-transportzone-672", "segmentation_id": 672, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf19d02ac-67", "ovs_interfaceid": "f19d02ac-6732-44b5-b454-df5777588a7e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1002.274269] env[63088]: DEBUG oslo_vmware.api [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Waiting for the task: (returnval){ [ 1002.274269] env[63088]: value = "task-1285402" [ 1002.274269] env[63088]: _type = "Task" [ 1002.274269] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.282727] env[63088]: DEBUG oslo_vmware.api [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285402, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.323037] env[63088]: DEBUG oslo_vmware.api [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285401, 'name': PowerOnVM_Task, 'duration_secs': 2.269933} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.323037] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1002.323037] env[63088]: DEBUG nova.compute.manager [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1002.323493] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-889c3158-798c-435a-bbbb-9f5513e83832 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.415623] env[63088]: DEBUG oslo_concurrency.lockutils [None req-818df68b-430f-49fa-83f0-07dc08efcfcd tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.786s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1002.422357] env[63088]: DEBUG oslo_concurrency.lockutils [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.839s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1002.422357] env[63088]: INFO nova.compute.claims [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1002.444304] env[63088]: INFO nova.scheduler.client.report [None req-818df68b-430f-49fa-83f0-07dc08efcfcd tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Deleted allocations for instance c4fb31b2-1301-4a6e-bef2-8ea462caee0d [ 1002.605487] env[63088]: DEBUG oslo_vmware.api [None req-4faf3ac2-8861-4d6a-8416-593c2f5689cf tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285389, 'name': ReconfigVM_Task} progress is 18%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.647749] env[63088]: DEBUG nova.compute.manager [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 32418785-8823-4a5e-90a9-e205752530b7] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1002.672437] env[63088]: DEBUG nova.virt.hardware [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1002.672703] env[63088]: DEBUG nova.virt.hardware [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1002.672896] env[63088]: DEBUG nova.virt.hardware [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1002.673115] env[63088]: DEBUG nova.virt.hardware [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1002.673272] env[63088]: DEBUG nova.virt.hardware [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1002.673422] env[63088]: DEBUG nova.virt.hardware [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1002.674056] env[63088]: DEBUG nova.virt.hardware [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1002.676021] env[63088]: DEBUG nova.virt.hardware [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1002.676021] env[63088]: DEBUG nova.virt.hardware [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1002.676021] env[63088]: DEBUG nova.virt.hardware [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1002.676021] env[63088]: DEBUG nova.virt.hardware [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1002.676021] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a515e9bc-cc27-4364-a763-ce6635a35462 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.685044] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f3bd226-dd4d-4c55-8a5b-3b4e5f46fb28 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.775518] env[63088]: DEBUG oslo_concurrency.lockutils [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Releasing lock "refresh_cache-a39840c1-3d60-478b-987c-2519551963cf" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1002.776032] env[63088]: DEBUG nova.compute.manager [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: a39840c1-3d60-478b-987c-2519551963cf] Instance network_info: |[{"id": "f19d02ac-6732-44b5-b454-df5777588a7e", "address": "fa:16:3e:14:a8:cd", "network": {"id": "dff14a7f-0af1-4e4e-a498-86d7c9816e6b", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-762073671-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f33f2701fad94864a8c406a404bc0a42", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "21310d90-efbc-45a8-a97f-c4358606530f", "external-id": "nsx-vlan-transportzone-672", "segmentation_id": 672, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf19d02ac-67", "ovs_interfaceid": "f19d02ac-6732-44b5-b454-df5777588a7e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1002.776647] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: a39840c1-3d60-478b-987c-2519551963cf] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:14:a8:cd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '21310d90-efbc-45a8-a97f-c4358606530f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f19d02ac-6732-44b5-b454-df5777588a7e', 'vif_model': 'vmxnet3'}] {{(pid=63088) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1002.784646] env[63088]: DEBUG oslo.service.loopingcall [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1002.788260] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a39840c1-3d60-478b-987c-2519551963cf] Creating VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1002.788512] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3d0742d4-f480-4bca-aa6e-217a39720e88 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.808422] env[63088]: DEBUG oslo_vmware.api [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285402, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.168359} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.809712] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1002.809900] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e3652b5e-39c5-4fe1-8b63-944e4527acd9/e3652b5e-39c5-4fe1-8b63-944e4527acd9.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1002.810167] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e3652b5e-39c5-4fe1-8b63-944e4527acd9/e3652b5e-39c5-4fe1-8b63-944e4527acd9.vmdk to [datastore2] 71e37d8e-a454-46c4-a3cc-3d5671a32beb/71e37d8e-a454-46c4-a3cc-3d5671a32beb.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1002.810592] env[63088]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1002.810592] env[63088]: value = "task-1285403" [ 1002.810592] env[63088]: _type = "Task" [ 1002.810592] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.810811] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a967cc77-3044-46b8-94a2-e4f80ab495c1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.820702] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285403, 'name': CreateVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.821989] env[63088]: DEBUG oslo_vmware.api [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Waiting for the task: (returnval){ [ 1002.821989] env[63088]: value = "task-1285404" [ 1002.821989] env[63088]: _type = "Task" [ 1002.821989] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.829671] env[63088]: DEBUG oslo_vmware.api [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285404, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.843327] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1002.952962] env[63088]: DEBUG oslo_concurrency.lockutils [None req-818df68b-430f-49fa-83f0-07dc08efcfcd tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lock "c4fb31b2-1301-4a6e-bef2-8ea462caee0d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.094s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1002.959499] env[63088]: DEBUG nova.compute.manager [req-3655efb2-ed6c-4b58-bb12-eb1118cf5573 req-c97c850b-c005-4269-83b6-f8f7a2091b21 service nova] [instance: a39840c1-3d60-478b-987c-2519551963cf] Received event network-changed-f19d02ac-6732-44b5-b454-df5777588a7e {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1002.959721] env[63088]: DEBUG nova.compute.manager [req-3655efb2-ed6c-4b58-bb12-eb1118cf5573 req-c97c850b-c005-4269-83b6-f8f7a2091b21 service nova] [instance: a39840c1-3d60-478b-987c-2519551963cf] Refreshing instance network info cache due to event network-changed-f19d02ac-6732-44b5-b454-df5777588a7e. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1002.959934] env[63088]: DEBUG oslo_concurrency.lockutils [req-3655efb2-ed6c-4b58-bb12-eb1118cf5573 req-c97c850b-c005-4269-83b6-f8f7a2091b21 service nova] Acquiring lock "refresh_cache-a39840c1-3d60-478b-987c-2519551963cf" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1002.960301] env[63088]: DEBUG oslo_concurrency.lockutils [req-3655efb2-ed6c-4b58-bb12-eb1118cf5573 req-c97c850b-c005-4269-83b6-f8f7a2091b21 service nova] Acquired lock "refresh_cache-a39840c1-3d60-478b-987c-2519551963cf" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1002.960567] env[63088]: DEBUG nova.network.neutron [req-3655efb2-ed6c-4b58-bb12-eb1118cf5573 req-c97c850b-c005-4269-83b6-f8f7a2091b21 service nova] [instance: a39840c1-3d60-478b-987c-2519551963cf] Refreshing network info cache for port f19d02ac-6732-44b5-b454-df5777588a7e {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1003.023154] env[63088]: DEBUG oslo_concurrency.lockutils [None req-e2510ea3-4fb6-4481-8321-61f1f1224284 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Acquiring lock "c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.023481] env[63088]: DEBUG oslo_concurrency.lockutils [None req-e2510ea3-4fb6-4481-8321-61f1f1224284 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Lock "c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.023708] env[63088]: DEBUG oslo_concurrency.lockutils [None req-e2510ea3-4fb6-4481-8321-61f1f1224284 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Acquiring lock "c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.023951] env[63088]: DEBUG oslo_concurrency.lockutils [None req-e2510ea3-4fb6-4481-8321-61f1f1224284 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Lock "c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.024263] env[63088]: DEBUG oslo_concurrency.lockutils [None req-e2510ea3-4fb6-4481-8321-61f1f1224284 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Lock "c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.030031] env[63088]: INFO nova.compute.manager [None req-e2510ea3-4fb6-4481-8321-61f1f1224284 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3] Terminating instance [ 1003.030031] env[63088]: DEBUG oslo_concurrency.lockutils [None req-e2510ea3-4fb6-4481-8321-61f1f1224284 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Acquiring lock "refresh_cache-c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1003.030031] env[63088]: DEBUG oslo_concurrency.lockutils [None req-e2510ea3-4fb6-4481-8321-61f1f1224284 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Acquired lock "refresh_cache-c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1003.030031] env[63088]: DEBUG nova.network.neutron [None req-e2510ea3-4fb6-4481-8321-61f1f1224284 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1003.037991] env[63088]: DEBUG nova.network.neutron [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 32418785-8823-4a5e-90a9-e205752530b7] Successfully updated port: 1e6d9172-35f9-4d1f-97ee-409ada1e1cd0 {{(pid=63088) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1003.104205] env[63088]: DEBUG oslo_vmware.api [None req-4faf3ac2-8861-4d6a-8416-593c2f5689cf tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285389, 'name': ReconfigVM_Task, 'duration_secs': 5.761883} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.104468] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4faf3ac2-8861-4d6a-8416-593c2f5689cf tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Releasing lock "d6fef3ef-fec8-4929-b9b8-5e63306aa51d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1003.104685] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-4faf3ac2-8861-4d6a-8416-593c2f5689cf tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Reconfigured VM to detach interface {{(pid=63088) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1984}} [ 1003.327064] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285403, 'name': CreateVM_Task} progress is 25%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.339636] env[63088]: DEBUG oslo_vmware.api [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285404, 'name': CopyVirtualDisk_Task} progress is 9%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.541460] env[63088]: DEBUG oslo_concurrency.lockutils [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquiring lock "refresh_cache-32418785-8823-4a5e-90a9-e205752530b7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1003.541660] env[63088]: DEBUG oslo_concurrency.lockutils [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquired lock "refresh_cache-32418785-8823-4a5e-90a9-e205752530b7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1003.541847] env[63088]: DEBUG nova.network.neutron [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 32418785-8823-4a5e-90a9-e205752530b7] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1003.558058] env[63088]: DEBUG nova.network.neutron [None req-e2510ea3-4fb6-4481-8321-61f1f1224284 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1003.610793] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a276821-2bec-4e77-9480-6f49c9f58ebf {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.624784] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05c15e7f-73c2-4b37-a20f-32f4e88313d8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.666393] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5db953e-86e6-401f-8c20-50e081442a02 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.675698] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dd76daf-7905-49cc-8331-224591cdc65e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.680778] env[63088]: DEBUG nova.network.neutron [None req-e2510ea3-4fb6-4481-8321-61f1f1224284 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1003.696873] env[63088]: DEBUG nova.compute.provider_tree [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1003.825025] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285403, 'name': CreateVM_Task, 'duration_secs': 0.658764} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.825264] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a39840c1-3d60-478b-987c-2519551963cf] Created VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1003.826058] env[63088]: DEBUG oslo_concurrency.lockutils [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1003.826299] env[63088]: DEBUG oslo_concurrency.lockutils [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1003.826676] env[63088]: DEBUG oslo_concurrency.lockutils [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1003.827040] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a27998c2-4641-4219-88af-a15ec156c33e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.837116] env[63088]: DEBUG oslo_vmware.api [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Waiting for the task: (returnval){ [ 1003.837116] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52d220e4-0674-f830-ea5e-dd5a49380c4d" [ 1003.837116] env[63088]: _type = "Task" [ 1003.837116] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.843100] env[63088]: DEBUG oslo_vmware.api [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285404, 'name': CopyVirtualDisk_Task} progress is 26%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.852462] env[63088]: DEBUG oslo_vmware.api [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52d220e4-0674-f830-ea5e-dd5a49380c4d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.853522] env[63088]: DEBUG nova.network.neutron [req-3655efb2-ed6c-4b58-bb12-eb1118cf5573 req-c97c850b-c005-4269-83b6-f8f7a2091b21 service nova] [instance: a39840c1-3d60-478b-987c-2519551963cf] Updated VIF entry in instance network info cache for port f19d02ac-6732-44b5-b454-df5777588a7e. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1003.853913] env[63088]: DEBUG nova.network.neutron [req-3655efb2-ed6c-4b58-bb12-eb1118cf5573 req-c97c850b-c005-4269-83b6-f8f7a2091b21 service nova] [instance: a39840c1-3d60-478b-987c-2519551963cf] Updating instance_info_cache with network_info: [{"id": "f19d02ac-6732-44b5-b454-df5777588a7e", "address": "fa:16:3e:14:a8:cd", "network": {"id": "dff14a7f-0af1-4e4e-a498-86d7c9816e6b", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-762073671-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f33f2701fad94864a8c406a404bc0a42", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "21310d90-efbc-45a8-a97f-c4358606530f", "external-id": "nsx-vlan-transportzone-672", "segmentation_id": 672, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf19d02ac-67", "ovs_interfaceid": "f19d02ac-6732-44b5-b454-df5777588a7e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1004.096313] env[63088]: DEBUG nova.network.neutron [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 32418785-8823-4a5e-90a9-e205752530b7] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1004.199237] env[63088]: DEBUG oslo_concurrency.lockutils [None req-e2510ea3-4fb6-4481-8321-61f1f1224284 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Releasing lock "refresh_cache-c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1004.199743] env[63088]: DEBUG nova.compute.manager [None req-e2510ea3-4fb6-4481-8321-61f1f1224284 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1004.199954] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-e2510ea3-4fb6-4481-8321-61f1f1224284 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1004.201049] env[63088]: DEBUG nova.scheduler.client.report [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1004.205124] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16bb99dd-d958-437a-863a-5fa28ddc045f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.215359] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2510ea3-4fb6-4481-8321-61f1f1224284 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1004.215655] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4cb769e3-91b4-4718-aa0e-a80f83992a99 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.225327] env[63088]: DEBUG oslo_vmware.api [None req-e2510ea3-4fb6-4481-8321-61f1f1224284 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Waiting for the task: (returnval){ [ 1004.225327] env[63088]: value = "task-1285405" [ 1004.225327] env[63088]: _type = "Task" [ 1004.225327] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.240530] env[63088]: DEBUG oslo_vmware.api [None req-e2510ea3-4fb6-4481-8321-61f1f1224284 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Task: {'id': task-1285405, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.309841] env[63088]: DEBUG oslo_concurrency.lockutils [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquiring lock "b3b1a32f-9cc1-4810-9984-fd24c40c81cd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.310152] env[63088]: DEBUG oslo_concurrency.lockutils [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lock "b3b1a32f-9cc1-4810-9984-fd24c40c81cd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1004.342624] env[63088]: DEBUG oslo_vmware.api [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285404, 'name': CopyVirtualDisk_Task} progress is 49%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.357226] env[63088]: DEBUG oslo_concurrency.lockutils [req-3655efb2-ed6c-4b58-bb12-eb1118cf5573 req-c97c850b-c005-4269-83b6-f8f7a2091b21 service nova] Releasing lock "refresh_cache-a39840c1-3d60-478b-987c-2519551963cf" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1004.357572] env[63088]: DEBUG nova.compute.manager [req-3655efb2-ed6c-4b58-bb12-eb1118cf5573 req-c97c850b-c005-4269-83b6-f8f7a2091b21 service nova] [instance: 32418785-8823-4a5e-90a9-e205752530b7] Received event network-vif-plugged-1e6d9172-35f9-4d1f-97ee-409ada1e1cd0 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1004.357785] env[63088]: DEBUG oslo_concurrency.lockutils [req-3655efb2-ed6c-4b58-bb12-eb1118cf5573 req-c97c850b-c005-4269-83b6-f8f7a2091b21 service nova] Acquiring lock "32418785-8823-4a5e-90a9-e205752530b7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.357996] env[63088]: DEBUG oslo_concurrency.lockutils [req-3655efb2-ed6c-4b58-bb12-eb1118cf5573 req-c97c850b-c005-4269-83b6-f8f7a2091b21 service nova] Lock "32418785-8823-4a5e-90a9-e205752530b7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1004.358177] env[63088]: DEBUG oslo_concurrency.lockutils [req-3655efb2-ed6c-4b58-bb12-eb1118cf5573 req-c97c850b-c005-4269-83b6-f8f7a2091b21 service nova] Lock "32418785-8823-4a5e-90a9-e205752530b7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1004.358402] env[63088]: DEBUG nova.compute.manager [req-3655efb2-ed6c-4b58-bb12-eb1118cf5573 req-c97c850b-c005-4269-83b6-f8f7a2091b21 service nova] [instance: 32418785-8823-4a5e-90a9-e205752530b7] No waiting events found dispatching network-vif-plugged-1e6d9172-35f9-4d1f-97ee-409ada1e1cd0 {{(pid=63088) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1004.358698] env[63088]: WARNING nova.compute.manager [req-3655efb2-ed6c-4b58-bb12-eb1118cf5573 req-c97c850b-c005-4269-83b6-f8f7a2091b21 service nova] [instance: 32418785-8823-4a5e-90a9-e205752530b7] Received unexpected event network-vif-plugged-1e6d9172-35f9-4d1f-97ee-409ada1e1cd0 for instance with vm_state building and task_state spawning. [ 1004.359073] env[63088]: DEBUG oslo_vmware.api [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52d220e4-0674-f830-ea5e-dd5a49380c4d, 'name': SearchDatastore_Task, 'duration_secs': 0.016492} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.359426] env[63088]: DEBUG oslo_concurrency.lockutils [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1004.359658] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: a39840c1-3d60-478b-987c-2519551963cf] Processing image 1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1004.363315] env[63088]: DEBUG oslo_concurrency.lockutils [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1004.363490] env[63088]: DEBUG oslo_concurrency.lockutils [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1004.363743] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1004.364036] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7e47b58d-95b3-4fd6-853f-d9fa71bd1e8f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.378126] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1004.378126] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1004.378126] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f0a5a3bf-029d-4357-924a-98f09109d346 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.384736] env[63088]: DEBUG oslo_vmware.api [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Waiting for the task: (returnval){ [ 1004.384736] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52e9917d-2f3d-cf62-fb4c-34f280b047bb" [ 1004.384736] env[63088]: _type = "Task" [ 1004.384736] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.393827] env[63088]: DEBUG oslo_vmware.api [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52e9917d-2f3d-cf62-fb4c-34f280b047bb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.441238] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f2b28811-7644-45df-ad39-abad0e2f8308 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquiring lock "d6fef3ef-fec8-4929-b9b8-5e63306aa51d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.441513] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f2b28811-7644-45df-ad39-abad0e2f8308 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Lock "d6fef3ef-fec8-4929-b9b8-5e63306aa51d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1004.441668] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f2b28811-7644-45df-ad39-abad0e2f8308 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquiring lock "d6fef3ef-fec8-4929-b9b8-5e63306aa51d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.441853] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f2b28811-7644-45df-ad39-abad0e2f8308 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Lock "d6fef3ef-fec8-4929-b9b8-5e63306aa51d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1004.442054] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f2b28811-7644-45df-ad39-abad0e2f8308 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Lock "d6fef3ef-fec8-4929-b9b8-5e63306aa51d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1004.444345] env[63088]: INFO nova.compute.manager [None req-f2b28811-7644-45df-ad39-abad0e2f8308 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Terminating instance [ 1004.446638] env[63088]: DEBUG nova.compute.manager [None req-f2b28811-7644-45df-ad39-abad0e2f8308 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1004.446805] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-f2b28811-7644-45df-ad39-abad0e2f8308 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1004.447663] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df3d9517-23e2-4d0b-b183-175c76aac6ae {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.457020] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2b28811-7644-45df-ad39-abad0e2f8308 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1004.457345] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cdb6528b-cec9-4178-ae8f-417d4327b4ca {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.466067] env[63088]: DEBUG oslo_vmware.api [None req-f2b28811-7644-45df-ad39-abad0e2f8308 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Waiting for the task: (returnval){ [ 1004.466067] env[63088]: value = "task-1285406" [ 1004.466067] env[63088]: _type = "Task" [ 1004.466067] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.476078] env[63088]: DEBUG oslo_vmware.api [None req-f2b28811-7644-45df-ad39-abad0e2f8308 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285406, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.480780] env[63088]: DEBUG nova.network.neutron [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 32418785-8823-4a5e-90a9-e205752530b7] Updating instance_info_cache with network_info: [{"id": "1e6d9172-35f9-4d1f-97ee-409ada1e1cd0", "address": "fa:16:3e:90:0e:bd", "network": {"id": "5b801381-6bb3-45cd-b81f-92b5d757e080", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1481457494-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e48f62dee9ad4e9b94b67bd871db5c63", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "051f343d-ac4f-4070-a26d-467603122c81", "external-id": "nsx-vlan-transportzone-277", "segmentation_id": 277, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1e6d9172-35", "ovs_interfaceid": "1e6d9172-35f9-4d1f-97ee-409ada1e1cd0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1004.503416] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4faf3ac2-8861-4d6a-8416-593c2f5689cf tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquiring lock "refresh_cache-d6fef3ef-fec8-4929-b9b8-5e63306aa51d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1004.503633] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4faf3ac2-8861-4d6a-8416-593c2f5689cf tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquired lock "refresh_cache-d6fef3ef-fec8-4929-b9b8-5e63306aa51d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1004.503919] env[63088]: DEBUG nova.network.neutron [None req-4faf3ac2-8861-4d6a-8416-593c2f5689cf tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1004.521219] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4043eead-2556-4461-8694-6718416fd3ce tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquiring lock "69645f50-46fc-4c15-9b39-1721f7636e31" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.521487] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4043eead-2556-4461-8694-6718416fd3ce tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lock "69645f50-46fc-4c15-9b39-1721f7636e31" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1004.523874] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4043eead-2556-4461-8694-6718416fd3ce tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquiring lock "69645f50-46fc-4c15-9b39-1721f7636e31-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.524105] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4043eead-2556-4461-8694-6718416fd3ce tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lock "69645f50-46fc-4c15-9b39-1721f7636e31-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1004.524347] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4043eead-2556-4461-8694-6718416fd3ce tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lock "69645f50-46fc-4c15-9b39-1721f7636e31-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1004.526566] env[63088]: INFO nova.compute.manager [None req-4043eead-2556-4461-8694-6718416fd3ce tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] Terminating instance [ 1004.528753] env[63088]: DEBUG nova.compute.manager [None req-4043eead-2556-4461-8694-6718416fd3ce tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1004.528973] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-4043eead-2556-4461-8694-6718416fd3ce tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1004.530154] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88450402-47fb-435e-adf3-65f631ece3b6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.538513] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-4043eead-2556-4461-8694-6718416fd3ce tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1004.538810] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-409e0797-c58d-47d8-a46c-433b30eac188 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.547414] env[63088]: DEBUG oslo_vmware.api [None req-4043eead-2556-4461-8694-6718416fd3ce tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for the task: (returnval){ [ 1004.547414] env[63088]: value = "task-1285407" [ 1004.547414] env[63088]: _type = "Task" [ 1004.547414] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.556550] env[63088]: DEBUG oslo_vmware.api [None req-4043eead-2556-4461-8694-6718416fd3ce tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285407, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.712163] env[63088]: DEBUG oslo_concurrency.lockutils [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.292s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1004.712163] env[63088]: DEBUG nova.compute.manager [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1004.714850] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 2.821s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1004.715146] env[63088]: DEBUG nova.objects.instance [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63088) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1004.736727] env[63088]: DEBUG oslo_vmware.api [None req-e2510ea3-4fb6-4481-8321-61f1f1224284 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Task: {'id': task-1285405, 'name': PowerOffVM_Task} progress is 100%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.816951] env[63088]: DEBUG nova.compute.manager [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: b3b1a32f-9cc1-4810-9984-fd24c40c81cd] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1004.843697] env[63088]: DEBUG oslo_vmware.api [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285404, 'name': CopyVirtualDisk_Task} progress is 69%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.898850] env[63088]: DEBUG oslo_vmware.api [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52e9917d-2f3d-cf62-fb4c-34f280b047bb, 'name': SearchDatastore_Task, 'duration_secs': 0.024268} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.900231] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-133d9cc5-7f5c-4c52-a5a3-3e7a7e1eb182 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.907974] env[63088]: DEBUG oslo_vmware.api [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Waiting for the task: (returnval){ [ 1004.907974] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52ed3189-a2d2-6424-881b-e0a6ba0114e0" [ 1004.907974] env[63088]: _type = "Task" [ 1004.907974] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.917370] env[63088]: DEBUG oslo_vmware.api [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52ed3189-a2d2-6424-881b-e0a6ba0114e0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.978117] env[63088]: DEBUG oslo_vmware.api [None req-f2b28811-7644-45df-ad39-abad0e2f8308 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285406, 'name': PowerOffVM_Task} progress is 100%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.984080] env[63088]: DEBUG oslo_concurrency.lockutils [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Releasing lock "refresh_cache-32418785-8823-4a5e-90a9-e205752530b7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1004.984333] env[63088]: DEBUG nova.compute.manager [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 32418785-8823-4a5e-90a9-e205752530b7] Instance network_info: |[{"id": "1e6d9172-35f9-4d1f-97ee-409ada1e1cd0", "address": "fa:16:3e:90:0e:bd", "network": {"id": "5b801381-6bb3-45cd-b81f-92b5d757e080", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1481457494-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e48f62dee9ad4e9b94b67bd871db5c63", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "051f343d-ac4f-4070-a26d-467603122c81", "external-id": "nsx-vlan-transportzone-277", "segmentation_id": 277, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1e6d9172-35", "ovs_interfaceid": "1e6d9172-35f9-4d1f-97ee-409ada1e1cd0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1004.984863] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 32418785-8823-4a5e-90a9-e205752530b7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:90:0e:bd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '051f343d-ac4f-4070-a26d-467603122c81', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1e6d9172-35f9-4d1f-97ee-409ada1e1cd0', 'vif_model': 'vmxnet3'}] {{(pid=63088) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1004.993075] env[63088]: DEBUG oslo.service.loopingcall [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1004.993773] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 32418785-8823-4a5e-90a9-e205752530b7] Creating VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1004.994112] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-acf2f726-7678-4f74-a957-0e46ba3bd20d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.013519] env[63088]: DEBUG nova.compute.manager [req-df1baea3-16e5-42db-9850-95e93a79144c req-770186b2-53fe-471d-9713-fb74ab92f96b service nova] [instance: 32418785-8823-4a5e-90a9-e205752530b7] Received event network-changed-1e6d9172-35f9-4d1f-97ee-409ada1e1cd0 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1005.013763] env[63088]: DEBUG nova.compute.manager [req-df1baea3-16e5-42db-9850-95e93a79144c req-770186b2-53fe-471d-9713-fb74ab92f96b service nova] [instance: 32418785-8823-4a5e-90a9-e205752530b7] Refreshing instance network info cache due to event network-changed-1e6d9172-35f9-4d1f-97ee-409ada1e1cd0. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1005.013966] env[63088]: DEBUG oslo_concurrency.lockutils [req-df1baea3-16e5-42db-9850-95e93a79144c req-770186b2-53fe-471d-9713-fb74ab92f96b service nova] Acquiring lock "refresh_cache-32418785-8823-4a5e-90a9-e205752530b7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1005.014196] env[63088]: DEBUG oslo_concurrency.lockutils [req-df1baea3-16e5-42db-9850-95e93a79144c req-770186b2-53fe-471d-9713-fb74ab92f96b service nova] Acquired lock "refresh_cache-32418785-8823-4a5e-90a9-e205752530b7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1005.014405] env[63088]: DEBUG nova.network.neutron [req-df1baea3-16e5-42db-9850-95e93a79144c req-770186b2-53fe-471d-9713-fb74ab92f96b service nova] [instance: 32418785-8823-4a5e-90a9-e205752530b7] Refreshing network info cache for port 1e6d9172-35f9-4d1f-97ee-409ada1e1cd0 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1005.024456] env[63088]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1005.024456] env[63088]: value = "task-1285408" [ 1005.024456] env[63088]: _type = "Task" [ 1005.024456] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.036578] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285408, 'name': CreateVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.058113] env[63088]: DEBUG oslo_vmware.api [None req-4043eead-2556-4461-8694-6718416fd3ce tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285407, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.221115] env[63088]: DEBUG nova.compute.utils [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1005.227093] env[63088]: DEBUG nova.compute.manager [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1005.227318] env[63088]: DEBUG nova.network.neutron [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1005.244933] env[63088]: DEBUG oslo_vmware.api [None req-e2510ea3-4fb6-4481-8321-61f1f1224284 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Task: {'id': task-1285405, 'name': PowerOffVM_Task, 'duration_secs': 0.5789} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.245448] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2510ea3-4fb6-4481-8321-61f1f1224284 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1005.245818] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-e2510ea3-4fb6-4481-8321-61f1f1224284 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3] Unregistering the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1005.246593] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-579db390-67b0-47e1-88ff-cdbdba12a638 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.253676] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1005.254192] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1005.285788] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-e2510ea3-4fb6-4481-8321-61f1f1224284 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3] Unregistered the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1005.286166] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-e2510ea3-4fb6-4481-8321-61f1f1224284 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3] Deleting contents of the VM from datastore datastore2 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1005.286419] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-e2510ea3-4fb6-4481-8321-61f1f1224284 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Deleting the datastore file [datastore2] c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3 {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1005.286798] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d620a77e-9467-4d80-b678-6f9e7a2743b0 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.298094] env[63088]: DEBUG oslo_vmware.api [None req-e2510ea3-4fb6-4481-8321-61f1f1224284 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Waiting for the task: (returnval){ [ 1005.298094] env[63088]: value = "task-1285410" [ 1005.298094] env[63088]: _type = "Task" [ 1005.298094] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.309994] env[63088]: DEBUG oslo_vmware.api [None req-e2510ea3-4fb6-4481-8321-61f1f1224284 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Task: {'id': task-1285410, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.311935] env[63088]: DEBUG nova.policy [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6ffb8b92c6d649c4b39d8df1c1d0c36c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a7898ddafe0d41038a6ae4277f2c2f48', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 1005.341476] env[63088]: DEBUG oslo_vmware.api [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285404, 'name': CopyVirtualDisk_Task} progress is 91%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.348440] env[63088]: DEBUG oslo_concurrency.lockutils [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1005.381682] env[63088]: INFO nova.network.neutron [None req-4faf3ac2-8861-4d6a-8416-593c2f5689cf tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Port 051455f3-4c94-4050-8610-bd8ace9f3988 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1005.382195] env[63088]: DEBUG nova.network.neutron [None req-4faf3ac2-8861-4d6a-8416-593c2f5689cf tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Updating instance_info_cache with network_info: [{"id": "b7a01048-9168-46ff-a5b8-83cee3dd3c97", "address": "fa:16:3e:e3:57:91", "network": {"id": "1174a284-d3ca-4f9e-aa81-13ee9a693e55", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1994276040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.150", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa3d24a1a6c0430985fd80365d986ee1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb7a01048-91", "ovs_interfaceid": "b7a01048-9168-46ff-a5b8-83cee3dd3c97", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1005.419459] env[63088]: DEBUG oslo_vmware.api [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52ed3189-a2d2-6424-881b-e0a6ba0114e0, 'name': SearchDatastore_Task, 'duration_secs': 0.025287} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.419750] env[63088]: DEBUG oslo_concurrency.lockutils [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1005.420066] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] a39840c1-3d60-478b-987c-2519551963cf/a39840c1-3d60-478b-987c-2519551963cf.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1005.420355] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1c07547b-ec89-4062-b7c6-6dfae4540ea2 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.428805] env[63088]: DEBUG oslo_vmware.api [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Waiting for the task: (returnval){ [ 1005.428805] env[63088]: value = "task-1285411" [ 1005.428805] env[63088]: _type = "Task" [ 1005.428805] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.438093] env[63088]: DEBUG oslo_vmware.api [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285411, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.479316] env[63088]: DEBUG oslo_vmware.api [None req-f2b28811-7644-45df-ad39-abad0e2f8308 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285406, 'name': PowerOffVM_Task, 'duration_secs': 0.518328} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.479545] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2b28811-7644-45df-ad39-abad0e2f8308 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1005.479610] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-f2b28811-7644-45df-ad39-abad0e2f8308 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Unregistering the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1005.479864] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9907a115-9897-4fad-9727-57995a8a6c74 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.537173] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285408, 'name': CreateVM_Task, 'duration_secs': 0.392322} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.537401] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 32418785-8823-4a5e-90a9-e205752530b7] Created VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1005.538186] env[63088]: DEBUG oslo_concurrency.lockutils [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1005.538410] env[63088]: DEBUG oslo_concurrency.lockutils [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1005.538897] env[63088]: DEBUG oslo_concurrency.lockutils [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1005.539206] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8dfa862d-79ad-4c10-a966-8aaad251409d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.545249] env[63088]: DEBUG oslo_vmware.api [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 1005.545249] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52fd998d-3216-edb6-83cc-f7e0d015120c" [ 1005.545249] env[63088]: _type = "Task" [ 1005.545249] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.559140] env[63088]: DEBUG oslo_vmware.api [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52fd998d-3216-edb6-83cc-f7e0d015120c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.563302] env[63088]: DEBUG oslo_vmware.api [None req-4043eead-2556-4461-8694-6718416fd3ce tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285407, 'name': PowerOffVM_Task} progress is 100%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.563670] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-f2b28811-7644-45df-ad39-abad0e2f8308 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Unregistered the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1005.563883] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-f2b28811-7644-45df-ad39-abad0e2f8308 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Deleting contents of the VM from datastore datastore2 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1005.564087] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-f2b28811-7644-45df-ad39-abad0e2f8308 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Deleting the datastore file [datastore2] d6fef3ef-fec8-4929-b9b8-5e63306aa51d {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1005.564364] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-38396fbf-8208-4890-8bee-48ed9bd385d5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.573095] env[63088]: DEBUG oslo_vmware.api [None req-f2b28811-7644-45df-ad39-abad0e2f8308 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Waiting for the task: (returnval){ [ 1005.573095] env[63088]: value = "task-1285413" [ 1005.573095] env[63088]: _type = "Task" [ 1005.573095] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.582047] env[63088]: DEBUG oslo_vmware.api [None req-f2b28811-7644-45df-ad39-abad0e2f8308 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285413, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.630847] env[63088]: DEBUG nova.network.neutron [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Successfully created port: f88388b7-bc70-4095-a792-8d155ef34534 {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1005.728152] env[63088]: DEBUG nova.compute.manager [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1005.735112] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a35afd25-d62a-4b6f-8218-2d477fc23560 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.018s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1005.735112] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 2.891s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1005.735112] env[63088]: DEBUG nova.objects.instance [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63088) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1005.738293] env[63088]: DEBUG nova.network.neutron [req-df1baea3-16e5-42db-9850-95e93a79144c req-770186b2-53fe-471d-9713-fb74ab92f96b service nova] [instance: 32418785-8823-4a5e-90a9-e205752530b7] Updated VIF entry in instance network info cache for port 1e6d9172-35f9-4d1f-97ee-409ada1e1cd0. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1005.739129] env[63088]: DEBUG nova.network.neutron [req-df1baea3-16e5-42db-9850-95e93a79144c req-770186b2-53fe-471d-9713-fb74ab92f96b service nova] [instance: 32418785-8823-4a5e-90a9-e205752530b7] Updating instance_info_cache with network_info: [{"id": "1e6d9172-35f9-4d1f-97ee-409ada1e1cd0", "address": "fa:16:3e:90:0e:bd", "network": {"id": "5b801381-6bb3-45cd-b81f-92b5d757e080", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1481457494-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e48f62dee9ad4e9b94b67bd871db5c63", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "051f343d-ac4f-4070-a26d-467603122c81", "external-id": "nsx-vlan-transportzone-277", "segmentation_id": 277, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1e6d9172-35", "ovs_interfaceid": "1e6d9172-35f9-4d1f-97ee-409ada1e1cd0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1005.763925] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1005.763925] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Starting heal instance info cache {{(pid=63088) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1005.810731] env[63088]: DEBUG oslo_vmware.api [None req-e2510ea3-4fb6-4481-8321-61f1f1224284 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Task: {'id': task-1285410, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.390875} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.811176] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-e2510ea3-4fb6-4481-8321-61f1f1224284 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1005.811295] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-e2510ea3-4fb6-4481-8321-61f1f1224284 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3] Deleted contents of the VM from datastore datastore2 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1005.811377] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-e2510ea3-4fb6-4481-8321-61f1f1224284 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1005.811558] env[63088]: INFO nova.compute.manager [None req-e2510ea3-4fb6-4481-8321-61f1f1224284 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3] Took 1.61 seconds to destroy the instance on the hypervisor. [ 1005.811840] env[63088]: DEBUG oslo.service.loopingcall [None req-e2510ea3-4fb6-4481-8321-61f1f1224284 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1005.812313] env[63088]: DEBUG nova.compute.manager [-] [instance: c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1005.812405] env[63088]: DEBUG nova.network.neutron [-] [instance: c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1005.829614] env[63088]: DEBUG nova.network.neutron [-] [instance: c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1005.844514] env[63088]: DEBUG oslo_vmware.api [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285404, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.744465} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.844795] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e3652b5e-39c5-4fe1-8b63-944e4527acd9/e3652b5e-39c5-4fe1-8b63-944e4527acd9.vmdk to [datastore2] 71e37d8e-a454-46c4-a3cc-3d5671a32beb/71e37d8e-a454-46c4-a3cc-3d5671a32beb.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1005.845630] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fecd7d6c-11b0-4349-b3b9-311c741b4af2 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.871688] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Reconfiguring VM instance instance-0000004d to attach disk [datastore2] 71e37d8e-a454-46c4-a3cc-3d5671a32beb/71e37d8e-a454-46c4-a3cc-3d5671a32beb.vmdk or device None with type streamOptimized {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1005.872409] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7494573e-459a-4d8c-89a7-f15217db184e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.888561] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4faf3ac2-8861-4d6a-8416-593c2f5689cf tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Releasing lock "refresh_cache-d6fef3ef-fec8-4929-b9b8-5e63306aa51d" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1005.896369] env[63088]: DEBUG oslo_vmware.api [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Waiting for the task: (returnval){ [ 1005.896369] env[63088]: value = "task-1285414" [ 1005.896369] env[63088]: _type = "Task" [ 1005.896369] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.907633] env[63088]: DEBUG oslo_vmware.api [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285414, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.941473] env[63088]: DEBUG oslo_vmware.api [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285411, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.059269] env[63088]: DEBUG oslo_vmware.api [None req-4043eead-2556-4461-8694-6718416fd3ce tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285407, 'name': PowerOffVM_Task, 'duration_secs': 1.041305} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.062916] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-4043eead-2556-4461-8694-6718416fd3ce tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1006.063159] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-4043eead-2556-4461-8694-6718416fd3ce tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] Unregistering the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1006.063441] env[63088]: DEBUG oslo_vmware.api [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52fd998d-3216-edb6-83cc-f7e0d015120c, 'name': SearchDatastore_Task, 'duration_secs': 0.026532} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.063639] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2ea95c76-ca09-4ced-9d49-7d56b6b0366e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.065354] env[63088]: DEBUG oslo_concurrency.lockutils [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1006.065608] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 32418785-8823-4a5e-90a9-e205752530b7] Processing image 1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1006.065854] env[63088]: DEBUG oslo_concurrency.lockutils [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1006.066034] env[63088]: DEBUG oslo_concurrency.lockutils [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1006.066248] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1006.066516] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a849e8f5-1bce-46b2-82dc-bb5b17448d8a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.078147] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1006.078351] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1006.079373] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4fc58bf7-963b-4d77-b7d9-955c536b23e6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.084530] env[63088]: DEBUG oslo_vmware.api [None req-f2b28811-7644-45df-ad39-abad0e2f8308 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285413, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.284426} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.085116] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-f2b28811-7644-45df-ad39-abad0e2f8308 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1006.085323] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-f2b28811-7644-45df-ad39-abad0e2f8308 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Deleted contents of the VM from datastore datastore2 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1006.085505] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-f2b28811-7644-45df-ad39-abad0e2f8308 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1006.085682] env[63088]: INFO nova.compute.manager [None req-f2b28811-7644-45df-ad39-abad0e2f8308 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Took 1.64 seconds to destroy the instance on the hypervisor. [ 1006.085923] env[63088]: DEBUG oslo.service.loopingcall [None req-f2b28811-7644-45df-ad39-abad0e2f8308 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1006.086137] env[63088]: DEBUG nova.compute.manager [-] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1006.086235] env[63088]: DEBUG nova.network.neutron [-] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1006.088765] env[63088]: DEBUG oslo_vmware.api [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 1006.088765] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]5206e1d2-a715-88af-a9d0-66e90ff748d4" [ 1006.088765] env[63088]: _type = "Task" [ 1006.088765] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.096630] env[63088]: DEBUG oslo_vmware.api [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]5206e1d2-a715-88af-a9d0-66e90ff748d4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.132677] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-4043eead-2556-4461-8694-6718416fd3ce tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] Unregistered the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1006.132935] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-4043eead-2556-4461-8694-6718416fd3ce tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] Deleting contents of the VM from datastore datastore2 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1006.133136] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-4043eead-2556-4461-8694-6718416fd3ce tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Deleting the datastore file [datastore2] 69645f50-46fc-4c15-9b39-1721f7636e31 {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1006.133418] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5b9a2596-3435-451f-b088-d226f85f5533 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.140230] env[63088]: DEBUG oslo_vmware.api [None req-4043eead-2556-4461-8694-6718416fd3ce tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for the task: (returnval){ [ 1006.140230] env[63088]: value = "task-1285416" [ 1006.140230] env[63088]: _type = "Task" [ 1006.140230] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.148541] env[63088]: DEBUG oslo_vmware.api [None req-4043eead-2556-4461-8694-6718416fd3ce tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285416, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.243487] env[63088]: DEBUG oslo_concurrency.lockutils [req-df1baea3-16e5-42db-9850-95e93a79144c req-770186b2-53fe-471d-9713-fb74ab92f96b service nova] Releasing lock "refresh_cache-32418785-8823-4a5e-90a9-e205752530b7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1006.339079] env[63088]: DEBUG nova.network.neutron [-] [instance: c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1006.393430] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4faf3ac2-8861-4d6a-8416-593c2f5689cf tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Lock "interface-d6fef3ef-fec8-4929-b9b8-5e63306aa51d-051455f3-4c94-4050-8610-bd8ace9f3988" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 10.653s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1006.406569] env[63088]: DEBUG oslo_vmware.api [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285414, 'name': ReconfigVM_Task, 'duration_secs': 0.419047} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.408115] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Reconfigured VM instance instance-0000004d to attach disk [datastore2] 71e37d8e-a454-46c4-a3cc-3d5671a32beb/71e37d8e-a454-46c4-a3cc-3d5671a32beb.vmdk or device None with type streamOptimized {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1006.409378] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'device_type': 'disk', 'device_name': '/dev/sda', 'encryption_options': None, 'size': 0, 'encryption_format': None, 'guest_format': None, 'encrypted': False, 'disk_bus': None, 'boot_index': 0, 'encryption_secret_uuid': None, 'image_id': '1e8c5d18-0a03-4e18-afe1-de5a6e255953'}], 'ephemerals': [], 'block_device_mapping': [{'device_type': None, 'guest_format': None, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-275932', 'volume_id': '8a71432a-16cf-4ea7-a27f-e1899ca9f9e8', 'name': 'volume-8a71432a-16cf-4ea7-a27f-e1899ca9f9e8', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '71e37d8e-a454-46c4-a3cc-3d5671a32beb', 'attached_at': '', 'detached_at': '', 'volume_id': '8a71432a-16cf-4ea7-a27f-e1899ca9f9e8', 'serial': '8a71432a-16cf-4ea7-a27f-e1899ca9f9e8'}, 'delete_on_termination': False, 'attachment_id': '8b4a6128-dde0-4cea-b0c2-86e6122a638a', 'disk_bus': None, 'boot_index': None, 'mount_device': '/dev/sdb', 'volume_type': None}], 'swap': None} {{(pid=63088) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1006.409782] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Volume attach. Driver type: vmdk {{(pid=63088) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1006.410124] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-275932', 'volume_id': '8a71432a-16cf-4ea7-a27f-e1899ca9f9e8', 'name': 'volume-8a71432a-16cf-4ea7-a27f-e1899ca9f9e8', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '71e37d8e-a454-46c4-a3cc-3d5671a32beb', 'attached_at': '', 'detached_at': '', 'volume_id': '8a71432a-16cf-4ea7-a27f-e1899ca9f9e8', 'serial': '8a71432a-16cf-4ea7-a27f-e1899ca9f9e8'} {{(pid=63088) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1006.412097] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8575250-3ef7-4524-8ecd-1199e2fbe537 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.437156] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-657b2637-5c31-4e29-9633-eef4012d2cf7 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.454034] env[63088]: DEBUG oslo_vmware.api [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285411, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.637599} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.467934] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] a39840c1-3d60-478b-987c-2519551963cf/a39840c1-3d60-478b-987c-2519551963cf.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1006.468435] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: a39840c1-3d60-478b-987c-2519551963cf] Extending root virtual disk to 1048576 {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1006.476463] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Reconfiguring VM instance instance-0000004d to attach disk [datastore1] volume-8a71432a-16cf-4ea7-a27f-e1899ca9f9e8/volume-8a71432a-16cf-4ea7-a27f-e1899ca9f9e8.vmdk or device None with type thin {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1006.478070] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3ebbcb52-697a-492c-8bcf-a1d7aac4c2e2 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.479114] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e9c130bf-98f6-4a06-a7c2-75b04bfca82d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.500023] env[63088]: DEBUG oslo_vmware.api [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Waiting for the task: (returnval){ [ 1006.500023] env[63088]: value = "task-1285417" [ 1006.500023] env[63088]: _type = "Task" [ 1006.500023] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.500023] env[63088]: DEBUG oslo_vmware.api [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Waiting for the task: (returnval){ [ 1006.500023] env[63088]: value = "task-1285418" [ 1006.500023] env[63088]: _type = "Task" [ 1006.500023] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.512877] env[63088]: DEBUG oslo_vmware.api [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285418, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.516381] env[63088]: DEBUG oslo_vmware.api [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285417, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.600673] env[63088]: DEBUG oslo_vmware.api [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]5206e1d2-a715-88af-a9d0-66e90ff748d4, 'name': SearchDatastore_Task, 'duration_secs': 0.010796} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.601323] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f9c44da9-b659-4e0e-a364-4cce711b33e5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.606885] env[63088]: DEBUG oslo_vmware.api [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 1006.606885] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52bf36b1-bdc5-9c69-bd5b-600365d46d53" [ 1006.606885] env[63088]: _type = "Task" [ 1006.606885] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.615216] env[63088]: DEBUG oslo_vmware.api [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52bf36b1-bdc5-9c69-bd5b-600365d46d53, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.650492] env[63088]: DEBUG oslo_vmware.api [None req-4043eead-2556-4461-8694-6718416fd3ce tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285416, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.199878} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.651017] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-4043eead-2556-4461-8694-6718416fd3ce tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1006.651232] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-4043eead-2556-4461-8694-6718416fd3ce tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] Deleted contents of the VM from datastore datastore2 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1006.651420] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-4043eead-2556-4461-8694-6718416fd3ce tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1006.651626] env[63088]: INFO nova.compute.manager [None req-4043eead-2556-4461-8694-6718416fd3ce tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] Took 2.12 seconds to destroy the instance on the hypervisor. [ 1006.651908] env[63088]: DEBUG oslo.service.loopingcall [None req-4043eead-2556-4461-8694-6718416fd3ce tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1006.652134] env[63088]: DEBUG nova.compute.manager [-] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1006.652232] env[63088]: DEBUG nova.network.neutron [-] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1006.745477] env[63088]: DEBUG nova.compute.manager [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1006.748536] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3d7a82be-758b-4d4c-a69b-df485e2f2cb0 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.014s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1006.751499] env[63088]: DEBUG oslo_concurrency.lockutils [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.402s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1006.756023] env[63088]: INFO nova.compute.claims [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: b3b1a32f-9cc1-4810-9984-fd24c40c81cd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1006.783252] env[63088]: DEBUG nova.virt.hardware [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1006.783564] env[63088]: DEBUG nova.virt.hardware [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1006.783735] env[63088]: DEBUG nova.virt.hardware [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1006.784050] env[63088]: DEBUG nova.virt.hardware [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1006.784253] env[63088]: DEBUG nova.virt.hardware [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1006.784426] env[63088]: DEBUG nova.virt.hardware [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1006.784653] env[63088]: DEBUG nova.virt.hardware [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1006.784853] env[63088]: DEBUG nova.virt.hardware [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1006.785078] env[63088]: DEBUG nova.virt.hardware [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1006.785264] env[63088]: DEBUG nova.virt.hardware [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1006.785449] env[63088]: DEBUG nova.virt.hardware [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1006.787466] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ef6263a-944e-4bc1-a898-a955906331da {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.798010] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b82fc69-a02c-4d41-b9ac-d0dc772ec1f1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.840395] env[63088]: INFO nova.compute.manager [-] [instance: c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3] Took 1.03 seconds to deallocate network for instance. [ 1007.016449] env[63088]: DEBUG oslo_vmware.api [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285417, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071038} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.022744] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: a39840c1-3d60-478b-987c-2519551963cf] Extended root virtual disk {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1007.023510] env[63088]: DEBUG oslo_vmware.api [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285418, 'name': ReconfigVM_Task, 'duration_secs': 0.33605} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.024295] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be879450-53fa-473f-b479-d3ec146c76e4 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.027210] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Reconfigured VM instance instance-0000004d to attach disk [datastore1] volume-8a71432a-16cf-4ea7-a27f-e1899ca9f9e8/volume-8a71432a-16cf-4ea7-a27f-e1899ca9f9e8.vmdk or device None with type thin {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1007.033284] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c98a88b2-dbc6-4e99-86d1-1f4c5dca7297 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.045847] env[63088]: DEBUG nova.compute.manager [req-e2dd8d5b-8ed5-44e9-814e-cb04cd03d18f req-5a88a9d3-f15e-4375-9948-b4bb774f5c25 service nova] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] Received event network-vif-deleted-23b3723c-8ff5-467b-b65b-7eb1ee341733 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1007.046186] env[63088]: INFO nova.compute.manager [req-e2dd8d5b-8ed5-44e9-814e-cb04cd03d18f req-5a88a9d3-f15e-4375-9948-b4bb774f5c25 service nova] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] Neutron deleted interface 23b3723c-8ff5-467b-b65b-7eb1ee341733; detaching it from the instance and deleting it from the info cache [ 1007.046380] env[63088]: DEBUG nova.network.neutron [req-e2dd8d5b-8ed5-44e9-814e-cb04cd03d18f req-5a88a9d3-f15e-4375-9948-b4bb774f5c25 service nova] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1007.074838] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: a39840c1-3d60-478b-987c-2519551963cf] Reconfiguring VM instance instance-00000061 to attach disk [datastore1] a39840c1-3d60-478b-987c-2519551963cf/a39840c1-3d60-478b-987c-2519551963cf.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1007.079734] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6b8acce6-67d3-4267-92ef-9c9bf97e7391 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.097969] env[63088]: DEBUG oslo_vmware.api [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Waiting for the task: (returnval){ [ 1007.097969] env[63088]: value = "task-1285419" [ 1007.097969] env[63088]: _type = "Task" [ 1007.097969] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.100283] env[63088]: DEBUG nova.compute.manager [req-8cd3927c-fe4f-4649-8bc5-e8eb165edcc3 req-f0788543-3aaf-4261-acab-0479765c5c66 service nova] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Received event network-vif-plugged-f88388b7-bc70-4095-a792-8d155ef34534 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1007.100749] env[63088]: DEBUG oslo_concurrency.lockutils [req-8cd3927c-fe4f-4649-8bc5-e8eb165edcc3 req-f0788543-3aaf-4261-acab-0479765c5c66 service nova] Acquiring lock "2b87a9f1-5f10-43c2-8bc9-6d560ec88015-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1007.100977] env[63088]: DEBUG oslo_concurrency.lockutils [req-8cd3927c-fe4f-4649-8bc5-e8eb165edcc3 req-f0788543-3aaf-4261-acab-0479765c5c66 service nova] Lock "2b87a9f1-5f10-43c2-8bc9-6d560ec88015-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.101172] env[63088]: DEBUG oslo_concurrency.lockutils [req-8cd3927c-fe4f-4649-8bc5-e8eb165edcc3 req-f0788543-3aaf-4261-acab-0479765c5c66 service nova] Lock "2b87a9f1-5f10-43c2-8bc9-6d560ec88015-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.101347] env[63088]: DEBUG nova.compute.manager [req-8cd3927c-fe4f-4649-8bc5-e8eb165edcc3 req-f0788543-3aaf-4261-acab-0479765c5c66 service nova] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] No waiting events found dispatching network-vif-plugged-f88388b7-bc70-4095-a792-8d155ef34534 {{(pid=63088) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1007.101786] env[63088]: WARNING nova.compute.manager [req-8cd3927c-fe4f-4649-8bc5-e8eb165edcc3 req-f0788543-3aaf-4261-acab-0479765c5c66 service nova] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Received unexpected event network-vif-plugged-f88388b7-bc70-4095-a792-8d155ef34534 for instance with vm_state building and task_state spawning. [ 1007.108983] env[63088]: DEBUG oslo_vmware.api [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Waiting for the task: (returnval){ [ 1007.108983] env[63088]: value = "task-1285420" [ 1007.108983] env[63088]: _type = "Task" [ 1007.108983] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.116977] env[63088]: DEBUG oslo_vmware.api [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285419, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.125445] env[63088]: DEBUG oslo_vmware.api [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285420, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.128962] env[63088]: DEBUG oslo_vmware.api [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52bf36b1-bdc5-9c69-bd5b-600365d46d53, 'name': SearchDatastore_Task, 'duration_secs': 0.011109} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.129226] env[63088]: DEBUG oslo_concurrency.lockutils [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1007.130251] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] 32418785-8823-4a5e-90a9-e205752530b7/32418785-8823-4a5e-90a9-e205752530b7.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1007.130251] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-dead4a15-21d5-466f-9aa3-7abbb675ad65 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.136477] env[63088]: DEBUG oslo_vmware.api [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 1007.136477] env[63088]: value = "task-1285421" [ 1007.136477] env[63088]: _type = "Task" [ 1007.136477] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.145472] env[63088]: DEBUG oslo_vmware.api [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285421, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.193186] env[63088]: DEBUG nova.network.neutron [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Successfully updated port: f88388b7-bc70-4095-a792-8d155ef34534 {{(pid=63088) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1007.272951] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Didn't find any instances for network info cache update. {{(pid=63088) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10012}} [ 1007.273233] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1007.273450] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1007.273656] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager.update_available_resource {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1007.347245] env[63088]: DEBUG oslo_concurrency.lockutils [None req-e2510ea3-4fb6-4481-8321-61f1f1224284 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1007.469021] env[63088]: DEBUG nova.network.neutron [-] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1007.550702] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-09477949-25da-40b4-91b3-2385cab2b44e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.561018] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66f974a0-6e74-4bb7-8325-db10a6a032ff {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.592194] env[63088]: DEBUG nova.compute.manager [req-e2dd8d5b-8ed5-44e9-814e-cb04cd03d18f req-5a88a9d3-f15e-4375-9948-b4bb774f5c25 service nova] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] Detach interface failed, port_id=23b3723c-8ff5-467b-b65b-7eb1ee341733, reason: Instance 69645f50-46fc-4c15-9b39-1721f7636e31 could not be found. {{(pid=63088) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1007.599890] env[63088]: DEBUG nova.network.neutron [-] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1007.613257] env[63088]: DEBUG oslo_vmware.api [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285419, 'name': ReconfigVM_Task, 'duration_secs': 0.16666} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.614405] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-275932', 'volume_id': '8a71432a-16cf-4ea7-a27f-e1899ca9f9e8', 'name': 'volume-8a71432a-16cf-4ea7-a27f-e1899ca9f9e8', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '71e37d8e-a454-46c4-a3cc-3d5671a32beb', 'attached_at': '', 'detached_at': '', 'volume_id': '8a71432a-16cf-4ea7-a27f-e1899ca9f9e8', 'serial': '8a71432a-16cf-4ea7-a27f-e1899ca9f9e8'} {{(pid=63088) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1007.615052] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cc05c537-6df5-4213-9f97-b9722faaeb9b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.628297] env[63088]: DEBUG oslo_vmware.api [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285420, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.629453] env[63088]: DEBUG oslo_vmware.api [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Waiting for the task: (returnval){ [ 1007.629453] env[63088]: value = "task-1285422" [ 1007.629453] env[63088]: _type = "Task" [ 1007.629453] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.637849] env[63088]: DEBUG oslo_vmware.api [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285422, 'name': Rename_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.648212] env[63088]: DEBUG oslo_vmware.api [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285421, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.508253} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.648508] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] 32418785-8823-4a5e-90a9-e205752530b7/32418785-8823-4a5e-90a9-e205752530b7.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1007.648742] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 32418785-8823-4a5e-90a9-e205752530b7] Extending root virtual disk to 1048576 {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1007.649077] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-39fe7103-ad26-4e66-9082-bdebd1b2f3a5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.657501] env[63088]: DEBUG oslo_vmware.api [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 1007.657501] env[63088]: value = "task-1285423" [ 1007.657501] env[63088]: _type = "Task" [ 1007.657501] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.667684] env[63088]: DEBUG oslo_vmware.api [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285423, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.696234] env[63088]: DEBUG oslo_concurrency.lockutils [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquiring lock "refresh_cache-2b87a9f1-5f10-43c2-8bc9-6d560ec88015" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1007.696380] env[63088]: DEBUG oslo_concurrency.lockutils [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquired lock "refresh_cache-2b87a9f1-5f10-43c2-8bc9-6d560ec88015" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1007.696528] env[63088]: DEBUG nova.network.neutron [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1007.776578] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1007.903695] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c8f364c-aca5-4d21-959e-7dd4bd9b2d6d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.911835] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9eca0d8-f225-47c1-a382-8ead4cfd26c8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.943357] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbf18238-7310-4d96-9aa8-70d186a11384 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.951329] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67aae60e-47c2-418b-8f29-ec9851eea811 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.966289] env[63088]: DEBUG nova.compute.provider_tree [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1007.971710] env[63088]: INFO nova.compute.manager [-] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] Took 1.32 seconds to deallocate network for instance. [ 1008.102495] env[63088]: INFO nova.compute.manager [-] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Took 2.02 seconds to deallocate network for instance. [ 1008.127033] env[63088]: DEBUG oslo_vmware.api [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285420, 'name': ReconfigVM_Task, 'duration_secs': 0.587604} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.127340] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: a39840c1-3d60-478b-987c-2519551963cf] Reconfigured VM instance instance-00000061 to attach disk [datastore1] a39840c1-3d60-478b-987c-2519551963cf/a39840c1-3d60-478b-987c-2519551963cf.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1008.127968] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-03b5a7ed-089e-4016-b455-64bfd01fc4d1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.138788] env[63088]: DEBUG oslo_vmware.api [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285422, 'name': Rename_Task, 'duration_secs': 0.162675} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.140068] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1008.140835] env[63088]: DEBUG oslo_vmware.api [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Waiting for the task: (returnval){ [ 1008.140835] env[63088]: value = "task-1285424" [ 1008.140835] env[63088]: _type = "Task" [ 1008.140835] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.141193] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-85ac2147-69df-44ef-a012-adf8511abcbc {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.151189] env[63088]: DEBUG oslo_vmware.api [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285424, 'name': Rename_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.153959] env[63088]: DEBUG oslo_vmware.api [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Waiting for the task: (returnval){ [ 1008.153959] env[63088]: value = "task-1285425" [ 1008.153959] env[63088]: _type = "Task" [ 1008.153959] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.164510] env[63088]: DEBUG oslo_vmware.api [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285425, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.170422] env[63088]: DEBUG oslo_vmware.api [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285423, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072842} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.170856] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 32418785-8823-4a5e-90a9-e205752530b7] Extended root virtual disk {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1008.171683] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7c32c1f-21f2-4735-9660-697ab8563eae {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.197509] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 32418785-8823-4a5e-90a9-e205752530b7] Reconfiguring VM instance instance-00000062 to attach disk [datastore1] 32418785-8823-4a5e-90a9-e205752530b7/32418785-8823-4a5e-90a9-e205752530b7.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1008.197929] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-62d95b4d-1caf-4f07-9c21-cc9aa47e3c1b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.221347] env[63088]: DEBUG oslo_vmware.api [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 1008.221347] env[63088]: value = "task-1285426" [ 1008.221347] env[63088]: _type = "Task" [ 1008.221347] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.230509] env[63088]: DEBUG oslo_vmware.api [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285426, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.243243] env[63088]: DEBUG nova.network.neutron [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1008.443313] env[63088]: DEBUG nova.network.neutron [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Updating instance_info_cache with network_info: [{"id": "f88388b7-bc70-4095-a792-8d155ef34534", "address": "fa:16:3e:f7:d1:c7", "network": {"id": "711d3d9a-86db-4aa1-87b5-481f1812dfa4", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-59376186-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a7898ddafe0d41038a6ae4277f2c2f48", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e41070eb-3ac1-4ca9-a3d0-fd65893a97de", "external-id": "nsx-vlan-transportzone-596", "segmentation_id": 596, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf88388b7-bc", "ovs_interfaceid": "f88388b7-bc70-4095-a792-8d155ef34534", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1008.469031] env[63088]: DEBUG nova.scheduler.client.report [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1008.477838] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4043eead-2556-4461-8694-6718416fd3ce tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1008.609681] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f2b28811-7644-45df-ad39-abad0e2f8308 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1008.653078] env[63088]: DEBUG oslo_vmware.api [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285424, 'name': Rename_Task, 'duration_secs': 0.181562} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.653366] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: a39840c1-3d60-478b-987c-2519551963cf] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1008.653618] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-613b2886-9878-445a-8ac5-8da251523de8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.664977] env[63088]: DEBUG oslo_vmware.api [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285425, 'name': PowerOnVM_Task, 'duration_secs': 0.496005} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.666255] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1008.668125] env[63088]: DEBUG oslo_vmware.api [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Waiting for the task: (returnval){ [ 1008.668125] env[63088]: value = "task-1285427" [ 1008.668125] env[63088]: _type = "Task" [ 1008.668125] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.676092] env[63088]: DEBUG oslo_vmware.api [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285427, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.731962] env[63088]: DEBUG oslo_vmware.api [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285426, 'name': ReconfigVM_Task, 'duration_secs': 0.334468} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.732376] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 32418785-8823-4a5e-90a9-e205752530b7] Reconfigured VM instance instance-00000062 to attach disk [datastore1] 32418785-8823-4a5e-90a9-e205752530b7/32418785-8823-4a5e-90a9-e205752530b7.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1008.733116] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dc7c5d86-6431-498a-8419-0f25db2ff27b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.743191] env[63088]: DEBUG oslo_vmware.api [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 1008.743191] env[63088]: value = "task-1285428" [ 1008.743191] env[63088]: _type = "Task" [ 1008.743191] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.752703] env[63088]: DEBUG oslo_vmware.api [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285428, 'name': Rename_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.769903] env[63088]: DEBUG nova.compute.manager [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1008.771486] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7327c4ed-d0e5-4b4f-b5f8-5069b4893d71 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.946744] env[63088]: DEBUG oslo_concurrency.lockutils [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Releasing lock "refresh_cache-2b87a9f1-5f10-43c2-8bc9-6d560ec88015" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1008.947263] env[63088]: DEBUG nova.compute.manager [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Instance network_info: |[{"id": "f88388b7-bc70-4095-a792-8d155ef34534", "address": "fa:16:3e:f7:d1:c7", "network": {"id": "711d3d9a-86db-4aa1-87b5-481f1812dfa4", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-59376186-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a7898ddafe0d41038a6ae4277f2c2f48", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e41070eb-3ac1-4ca9-a3d0-fd65893a97de", "external-id": "nsx-vlan-transportzone-596", "segmentation_id": 596, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf88388b7-bc", "ovs_interfaceid": "f88388b7-bc70-4095-a792-8d155ef34534", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1008.948303] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f7:d1:c7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e41070eb-3ac1-4ca9-a3d0-fd65893a97de', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f88388b7-bc70-4095-a792-8d155ef34534', 'vif_model': 'vmxnet3'}] {{(pid=63088) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1008.957819] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Creating folder: Project (a7898ddafe0d41038a6ae4277f2c2f48). Parent ref: group-v275816. {{(pid=63088) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1008.958224] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-509cabb3-d5ed-4b6a-8473-16fc771bb084 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.973800] env[63088]: DEBUG oslo_concurrency.lockutils [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.223s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1008.974497] env[63088]: DEBUG nova.compute.manager [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: b3b1a32f-9cc1-4810-9984-fd24c40c81cd] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1008.978490] env[63088]: DEBUG oslo_concurrency.lockutils [None req-e2510ea3-4fb6-4481-8321-61f1f1224284 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.631s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.978771] env[63088]: DEBUG nova.objects.instance [None req-e2510ea3-4fb6-4481-8321-61f1f1224284 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Lazy-loading 'resources' on Instance uuid c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1008.980059] env[63088]: INFO nova.virt.vmwareapi.vm_util [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Created folder: Project (a7898ddafe0d41038a6ae4277f2c2f48) in parent group-v275816. [ 1008.980233] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Creating folder: Instances. Parent ref: group-v275947. {{(pid=63088) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1008.980799] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-83e8552e-2f1d-426a-89f0-b45e5c1e7c1e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.992193] env[63088]: INFO nova.virt.vmwareapi.vm_util [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Created folder: Instances in parent group-v275947. [ 1008.992477] env[63088]: DEBUG oslo.service.loopingcall [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1008.993174] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Creating VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1008.993174] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2aca2698-c44e-4011-994c-e6b0050ed557 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.015145] env[63088]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1009.015145] env[63088]: value = "task-1285431" [ 1009.015145] env[63088]: _type = "Task" [ 1009.015145] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.024185] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285431, 'name': CreateVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.058529] env[63088]: DEBUG nova.compute.manager [req-61ac1761-7825-4862-a599-92982e95a288 req-3b46d113-dcc9-4dda-8a3d-9299bcf1d192 service nova] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Received event network-vif-deleted-b7a01048-9168-46ff-a5b8-83cee3dd3c97 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1009.112972] env[63088]: DEBUG nova.compute.manager [req-88cdabcd-727b-409e-8123-6e0bad2f8af3 req-326a1512-ca83-409c-8e2d-d84a12b130b6 service nova] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Received event network-changed-f88388b7-bc70-4095-a792-8d155ef34534 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1009.113246] env[63088]: DEBUG nova.compute.manager [req-88cdabcd-727b-409e-8123-6e0bad2f8af3 req-326a1512-ca83-409c-8e2d-d84a12b130b6 service nova] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Refreshing instance network info cache due to event network-changed-f88388b7-bc70-4095-a792-8d155ef34534. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1009.113492] env[63088]: DEBUG oslo_concurrency.lockutils [req-88cdabcd-727b-409e-8123-6e0bad2f8af3 req-326a1512-ca83-409c-8e2d-d84a12b130b6 service nova] Acquiring lock "refresh_cache-2b87a9f1-5f10-43c2-8bc9-6d560ec88015" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1009.113662] env[63088]: DEBUG oslo_concurrency.lockutils [req-88cdabcd-727b-409e-8123-6e0bad2f8af3 req-326a1512-ca83-409c-8e2d-d84a12b130b6 service nova] Acquired lock "refresh_cache-2b87a9f1-5f10-43c2-8bc9-6d560ec88015" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1009.113857] env[63088]: DEBUG nova.network.neutron [req-88cdabcd-727b-409e-8123-6e0bad2f8af3 req-326a1512-ca83-409c-8e2d-d84a12b130b6 service nova] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Refreshing network info cache for port f88388b7-bc70-4095-a792-8d155ef34534 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1009.178548] env[63088]: DEBUG oslo_vmware.api [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285427, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.254751] env[63088]: DEBUG oslo_vmware.api [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285428, 'name': Rename_Task, 'duration_secs': 0.323231} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.255070] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 32418785-8823-4a5e-90a9-e205752530b7] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1009.255358] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-39d9c40a-c3b4-411d-9325-0c2cacbd805c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.273991] env[63088]: DEBUG oslo_vmware.api [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 1009.273991] env[63088]: value = "task-1285432" [ 1009.273991] env[63088]: _type = "Task" [ 1009.273991] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.287187] env[63088]: DEBUG oslo_vmware.api [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285432, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.289829] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b144ed78-b180-4554-968e-a46a47faeeef tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Lock "71e37d8e-a454-46c4-a3cc-3d5671a32beb" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 34.513s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.485748] env[63088]: DEBUG nova.compute.utils [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1009.491248] env[63088]: DEBUG nova.compute.manager [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: b3b1a32f-9cc1-4810-9984-fd24c40c81cd] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1009.491417] env[63088]: DEBUG nova.network.neutron [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: b3b1a32f-9cc1-4810-9984-fd24c40c81cd] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1009.530188] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285431, 'name': CreateVM_Task} progress is 99%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.536135] env[63088]: DEBUG nova.policy [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '32d5ff9b3ea345e4ac56010aa917e51b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '03e20d606e654362acbe2b36fe499ae3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 1009.660996] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-114319c7-c25a-47f2-84f2-4002df1e9676 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.675852] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60093153-e7e1-48d0-8ace-ec76a3e55541 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.685523] env[63088]: DEBUG oslo_vmware.api [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285427, 'name': PowerOnVM_Task} progress is 78%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.713036] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c46dd046-b454-4b1a-a776-260f9a5807ca {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.721626] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eaa02c80-667b-41b7-84ef-9ffd8c5b9bab {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.736653] env[63088]: DEBUG nova.compute.provider_tree [None req-e2510ea3-4fb6-4481-8321-61f1f1224284 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1009.785525] env[63088]: DEBUG oslo_vmware.api [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285432, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.854698] env[63088]: DEBUG nova.network.neutron [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: b3b1a32f-9cc1-4810-9984-fd24c40c81cd] Successfully created port: 36ea0e72-8934-47fa-8982-6b888b47766e {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1009.864689] env[63088]: DEBUG nova.network.neutron [req-88cdabcd-727b-409e-8123-6e0bad2f8af3 req-326a1512-ca83-409c-8e2d-d84a12b130b6 service nova] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Updated VIF entry in instance network info cache for port f88388b7-bc70-4095-a792-8d155ef34534. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1009.865093] env[63088]: DEBUG nova.network.neutron [req-88cdabcd-727b-409e-8123-6e0bad2f8af3 req-326a1512-ca83-409c-8e2d-d84a12b130b6 service nova] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Updating instance_info_cache with network_info: [{"id": "f88388b7-bc70-4095-a792-8d155ef34534", "address": "fa:16:3e:f7:d1:c7", "network": {"id": "711d3d9a-86db-4aa1-87b5-481f1812dfa4", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-59376186-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a7898ddafe0d41038a6ae4277f2c2f48", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e41070eb-3ac1-4ca9-a3d0-fd65893a97de", "external-id": "nsx-vlan-transportzone-596", "segmentation_id": 596, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf88388b7-bc", "ovs_interfaceid": "f88388b7-bc70-4095-a792-8d155ef34534", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1009.992115] env[63088]: DEBUG nova.compute.manager [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: b3b1a32f-9cc1-4810-9984-fd24c40c81cd] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1010.030092] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285431, 'name': CreateVM_Task, 'duration_secs': 0.538892} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.030345] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Created VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1010.031306] env[63088]: DEBUG oslo_concurrency.lockutils [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1010.031540] env[63088]: DEBUG oslo_concurrency.lockutils [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1010.031878] env[63088]: DEBUG oslo_concurrency.lockutils [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1010.032190] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cffc2dbb-33dd-459c-addb-e25d5791aec3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.037682] env[63088]: DEBUG oslo_vmware.api [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1010.037682] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52723b3d-0515-7d37-0943-4d1e2f7295ab" [ 1010.037682] env[63088]: _type = "Task" [ 1010.037682] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.046500] env[63088]: DEBUG oslo_vmware.api [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52723b3d-0515-7d37-0943-4d1e2f7295ab, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.182010] env[63088]: DEBUG oslo_vmware.api [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285427, 'name': PowerOnVM_Task, 'duration_secs': 1.109986} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.182309] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: a39840c1-3d60-478b-987c-2519551963cf] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1010.182519] env[63088]: INFO nova.compute.manager [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: a39840c1-3d60-478b-987c-2519551963cf] Took 9.84 seconds to spawn the instance on the hypervisor. [ 1010.182789] env[63088]: DEBUG nova.compute.manager [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: a39840c1-3d60-478b-987c-2519551963cf] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1010.183576] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-511fa9a2-1d44-4cf2-a4d3-95eb8664e44e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.240952] env[63088]: DEBUG nova.scheduler.client.report [None req-e2510ea3-4fb6-4481-8321-61f1f1224284 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1010.284840] env[63088]: DEBUG oslo_vmware.api [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285432, 'name': PowerOnVM_Task, 'duration_secs': 0.540616} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.285125] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 32418785-8823-4a5e-90a9-e205752530b7] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1010.285331] env[63088]: INFO nova.compute.manager [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 32418785-8823-4a5e-90a9-e205752530b7] Took 7.64 seconds to spawn the instance on the hypervisor. [ 1010.285561] env[63088]: DEBUG nova.compute.manager [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 32418785-8823-4a5e-90a9-e205752530b7] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1010.286407] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6fb7cb4-e4be-470a-b3ad-e4c27d3cf6fe {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.369080] env[63088]: DEBUG oslo_concurrency.lockutils [req-88cdabcd-727b-409e-8123-6e0bad2f8af3 req-326a1512-ca83-409c-8e2d-d84a12b130b6 service nova] Releasing lock "refresh_cache-2b87a9f1-5f10-43c2-8bc9-6d560ec88015" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1010.552421] env[63088]: DEBUG oslo_vmware.api [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52723b3d-0515-7d37-0943-4d1e2f7295ab, 'name': SearchDatastore_Task, 'duration_secs': 0.013157} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.552744] env[63088]: DEBUG oslo_concurrency.lockutils [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1010.552982] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Processing image 1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1010.553280] env[63088]: DEBUG oslo_concurrency.lockutils [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1010.553363] env[63088]: DEBUG oslo_concurrency.lockutils [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1010.553544] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1010.553802] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-832d3ab2-8157-46dd-9d78-c27af0f3056f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.563084] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1010.563275] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1010.563971] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ce605585-2be2-4ce7-aa1e-f38622255db2 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.570095] env[63088]: DEBUG oslo_vmware.api [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1010.570095] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]5282cd51-92d6-b5be-122a-82f205f6b279" [ 1010.570095] env[63088]: _type = "Task" [ 1010.570095] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.578125] env[63088]: DEBUG oslo_vmware.api [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]5282cd51-92d6-b5be-122a-82f205f6b279, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.701773] env[63088]: INFO nova.compute.manager [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: a39840c1-3d60-478b-987c-2519551963cf] Took 22.07 seconds to build instance. [ 1010.745647] env[63088]: DEBUG oslo_concurrency.lockutils [None req-e2510ea3-4fb6-4481-8321-61f1f1224284 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.767s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.748860] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 2.972s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.748860] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.748860] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63088) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1010.749019] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4043eead-2556-4461-8694-6718416fd3ce tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.272s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.749201] env[63088]: DEBUG nova.objects.instance [None req-4043eead-2556-4461-8694-6718416fd3ce tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lazy-loading 'resources' on Instance uuid 69645f50-46fc-4c15-9b39-1721f7636e31 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1010.751149] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0abff207-64f7-4aeb-a658-8fcc6035afba {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.761160] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32c29f1b-4f20-4093-91f1-77750be6dfbc {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.779947] env[63088]: INFO nova.scheduler.client.report [None req-e2510ea3-4fb6-4481-8321-61f1f1224284 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Deleted allocations for instance c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3 [ 1010.780766] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9634a5d1-8788-4169-832d-f5a479d21812 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.791035] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa4f8067-3d32-4b22-bef4-6d3456b4468c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.825335] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180701MB free_disk=140GB free_vcpus=48 pci_devices=None {{(pid=63088) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1010.825552] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.829673] env[63088]: INFO nova.compute.manager [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 32418785-8823-4a5e-90a9-e205752530b7] Took 21.33 seconds to build instance. [ 1011.002028] env[63088]: DEBUG nova.compute.manager [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: b3b1a32f-9cc1-4810-9984-fd24c40c81cd] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1011.032207] env[63088]: DEBUG nova.virt.hardware [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1011.032472] env[63088]: DEBUG nova.virt.hardware [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1011.032636] env[63088]: DEBUG nova.virt.hardware [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1011.032826] env[63088]: DEBUG nova.virt.hardware [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1011.032978] env[63088]: DEBUG nova.virt.hardware [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1011.033141] env[63088]: DEBUG nova.virt.hardware [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1011.033351] env[63088]: DEBUG nova.virt.hardware [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1011.033516] env[63088]: DEBUG nova.virt.hardware [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1011.033689] env[63088]: DEBUG nova.virt.hardware [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1011.033861] env[63088]: DEBUG nova.virt.hardware [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1011.034039] env[63088]: DEBUG nova.virt.hardware [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1011.034911] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee9a8a76-6133-498f-a664-e443c3bd0f2e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.043708] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea91f75f-2e64-4ff7-8c14-e4f739847508 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.050479] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae5dccb1-c85d-4e95-af86-9c20653519db {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.065431] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-98725594-259f-434a-88a1-810b26a94db4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 32418785-8823-4a5e-90a9-e205752530b7] Suspending the VM {{(pid=63088) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1163}} [ 1011.065691] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-c824f0c0-eab1-4b57-a963-2e33c95dcf13 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.074538] env[63088]: DEBUG oslo_vmware.api [None req-98725594-259f-434a-88a1-810b26a94db4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 1011.074538] env[63088]: value = "task-1285433" [ 1011.074538] env[63088]: _type = "Task" [ 1011.074538] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.081788] env[63088]: DEBUG oslo_vmware.api [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]5282cd51-92d6-b5be-122a-82f205f6b279, 'name': SearchDatastore_Task, 'duration_secs': 0.01961} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.082989] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5a8c886a-1208-4a10-a7dc-f5826722a284 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.089988] env[63088]: DEBUG oslo_vmware.api [None req-98725594-259f-434a-88a1-810b26a94db4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285433, 'name': SuspendVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.093723] env[63088]: DEBUG oslo_vmware.api [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1011.093723] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]5264da67-e54d-7554-5a82-f72a7cfef464" [ 1011.093723] env[63088]: _type = "Task" [ 1011.093723] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.101819] env[63088]: DEBUG oslo_vmware.api [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]5264da67-e54d-7554-5a82-f72a7cfef464, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.203646] env[63088]: DEBUG oslo_concurrency.lockutils [None req-27b623e2-2c7d-457a-b8e1-28e0f82520b7 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Lock "a39840c1-3d60-478b-987c-2519551963cf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.578s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1011.288708] env[63088]: DEBUG oslo_concurrency.lockutils [None req-e2510ea3-4fb6-4481-8321-61f1f1224284 tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Lock "c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.265s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1011.341910] env[63088]: DEBUG oslo_concurrency.lockutils [None req-fabec891-007f-46e0-be45-2aa8b4b13057 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "32418785-8823-4a5e-90a9-e205752530b7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.849s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1011.373360] env[63088]: DEBUG nova.compute.manager [req-b4a7d9a3-fcd3-47e1-8f65-731c89ebafaf req-e5790d06-3ea6-4c00-a951-9fdfca5bf967 service nova] [instance: b3b1a32f-9cc1-4810-9984-fd24c40c81cd] Received event network-vif-plugged-36ea0e72-8934-47fa-8982-6b888b47766e {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1011.373596] env[63088]: DEBUG oslo_concurrency.lockutils [req-b4a7d9a3-fcd3-47e1-8f65-731c89ebafaf req-e5790d06-3ea6-4c00-a951-9fdfca5bf967 service nova] Acquiring lock "b3b1a32f-9cc1-4810-9984-fd24c40c81cd-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.373871] env[63088]: DEBUG oslo_concurrency.lockutils [req-b4a7d9a3-fcd3-47e1-8f65-731c89ebafaf req-e5790d06-3ea6-4c00-a951-9fdfca5bf967 service nova] Lock "b3b1a32f-9cc1-4810-9984-fd24c40c81cd-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.374522] env[63088]: DEBUG oslo_concurrency.lockutils [req-b4a7d9a3-fcd3-47e1-8f65-731c89ebafaf req-e5790d06-3ea6-4c00-a951-9fdfca5bf967 service nova] Lock "b3b1a32f-9cc1-4810-9984-fd24c40c81cd-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1011.374790] env[63088]: DEBUG nova.compute.manager [req-b4a7d9a3-fcd3-47e1-8f65-731c89ebafaf req-e5790d06-3ea6-4c00-a951-9fdfca5bf967 service nova] [instance: b3b1a32f-9cc1-4810-9984-fd24c40c81cd] No waiting events found dispatching network-vif-plugged-36ea0e72-8934-47fa-8982-6b888b47766e {{(pid=63088) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1011.374989] env[63088]: WARNING nova.compute.manager [req-b4a7d9a3-fcd3-47e1-8f65-731c89ebafaf req-e5790d06-3ea6-4c00-a951-9fdfca5bf967 service nova] [instance: b3b1a32f-9cc1-4810-9984-fd24c40c81cd] Received unexpected event network-vif-plugged-36ea0e72-8934-47fa-8982-6b888b47766e for instance with vm_state building and task_state spawning. [ 1011.425309] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-719f9f69-6878-4a86-b8b0-46fc7b9995eb {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.437164] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9e2dd12-3007-4933-ae0e-f4c23b79630b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.470704] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d14d2b0-09f8-4257-8a96-fbccd76da410 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.479018] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0127e28e-2630-4f34-bcaa-626db3a6fbf0 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.495887] env[63088]: DEBUG nova.compute.provider_tree [None req-4043eead-2556-4461-8694-6718416fd3ce tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1011.495887] env[63088]: DEBUG nova.network.neutron [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: b3b1a32f-9cc1-4810-9984-fd24c40c81cd] Successfully updated port: 36ea0e72-8934-47fa-8982-6b888b47766e {{(pid=63088) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1011.585701] env[63088]: DEBUG oslo_vmware.api [None req-98725594-259f-434a-88a1-810b26a94db4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285433, 'name': SuspendVM_Task} progress is 58%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.610123] env[63088]: DEBUG oslo_vmware.api [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]5264da67-e54d-7554-5a82-f72a7cfef464, 'name': SearchDatastore_Task, 'duration_secs': 0.011369} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.610123] env[63088]: DEBUG oslo_concurrency.lockutils [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1011.610123] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] 2b87a9f1-5f10-43c2-8bc9-6d560ec88015/2b87a9f1-5f10-43c2-8bc9-6d560ec88015.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1011.610374] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-45821ba0-f8fa-4026-84cc-6036176cc3d8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.619161] env[63088]: DEBUG oslo_vmware.api [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1011.619161] env[63088]: value = "task-1285434" [ 1011.619161] env[63088]: _type = "Task" [ 1011.619161] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.628251] env[63088]: DEBUG oslo_vmware.api [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285434, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.845748] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5fefe597-a5a6-4e93-89ab-253f7983dd2f tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Acquiring lock "13388466-4adb-4d56-9fc1-e3f5dc516077" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.846055] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5fefe597-a5a6-4e93-89ab-253f7983dd2f tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Lock "13388466-4adb-4d56-9fc1-e3f5dc516077" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.846280] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5fefe597-a5a6-4e93-89ab-253f7983dd2f tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Acquiring lock "13388466-4adb-4d56-9fc1-e3f5dc516077-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.846469] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5fefe597-a5a6-4e93-89ab-253f7983dd2f tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Lock "13388466-4adb-4d56-9fc1-e3f5dc516077-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.846640] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5fefe597-a5a6-4e93-89ab-253f7983dd2f tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Lock "13388466-4adb-4d56-9fc1-e3f5dc516077-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1011.851479] env[63088]: INFO nova.compute.manager [None req-5fefe597-a5a6-4e93-89ab-253f7983dd2f tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: 13388466-4adb-4d56-9fc1-e3f5dc516077] Terminating instance [ 1011.852727] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5fefe597-a5a6-4e93-89ab-253f7983dd2f tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Acquiring lock "refresh_cache-13388466-4adb-4d56-9fc1-e3f5dc516077" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1011.852891] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5fefe597-a5a6-4e93-89ab-253f7983dd2f tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Acquired lock "refresh_cache-13388466-4adb-4d56-9fc1-e3f5dc516077" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1011.853205] env[63088]: DEBUG nova.network.neutron [None req-5fefe597-a5a6-4e93-89ab-253f7983dd2f tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: 13388466-4adb-4d56-9fc1-e3f5dc516077] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1011.998082] env[63088]: DEBUG nova.scheduler.client.report [None req-4043eead-2556-4461-8694-6718416fd3ce tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1012.001865] env[63088]: DEBUG oslo_concurrency.lockutils [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquiring lock "refresh_cache-b3b1a32f-9cc1-4810-9984-fd24c40c81cd" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1012.002020] env[63088]: DEBUG oslo_concurrency.lockutils [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquired lock "refresh_cache-b3b1a32f-9cc1-4810-9984-fd24c40c81cd" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1012.002191] env[63088]: DEBUG nova.network.neutron [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: b3b1a32f-9cc1-4810-9984-fd24c40c81cd] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1012.087215] env[63088]: DEBUG oslo_vmware.api [None req-98725594-259f-434a-88a1-810b26a94db4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285433, 'name': SuspendVM_Task, 'duration_secs': 0.634476} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.087528] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-98725594-259f-434a-88a1-810b26a94db4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 32418785-8823-4a5e-90a9-e205752530b7] Suspended the VM {{(pid=63088) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1167}} [ 1012.087722] env[63088]: DEBUG nova.compute.manager [None req-98725594-259f-434a-88a1-810b26a94db4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 32418785-8823-4a5e-90a9-e205752530b7] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1012.088549] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f85381c-c027-4fff-a80b-6ee4568fe4d1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.104224] env[63088]: DEBUG oslo_concurrency.lockutils [None req-407352de-05ec-47cb-9bee-a7de5aaee2c3 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquiring lock "a39840c1-3d60-478b-987c-2519551963cf" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1012.104503] env[63088]: DEBUG oslo_concurrency.lockutils [None req-407352de-05ec-47cb-9bee-a7de5aaee2c3 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Lock "a39840c1-3d60-478b-987c-2519551963cf" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1012.104877] env[63088]: DEBUG oslo_concurrency.lockutils [None req-407352de-05ec-47cb-9bee-a7de5aaee2c3 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquiring lock "a39840c1-3d60-478b-987c-2519551963cf-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1012.104925] env[63088]: DEBUG oslo_concurrency.lockutils [None req-407352de-05ec-47cb-9bee-a7de5aaee2c3 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Lock "a39840c1-3d60-478b-987c-2519551963cf-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1012.105093] env[63088]: DEBUG oslo_concurrency.lockutils [None req-407352de-05ec-47cb-9bee-a7de5aaee2c3 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Lock "a39840c1-3d60-478b-987c-2519551963cf-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1012.107698] env[63088]: INFO nova.compute.manager [None req-407352de-05ec-47cb-9bee-a7de5aaee2c3 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: a39840c1-3d60-478b-987c-2519551963cf] Terminating instance [ 1012.111161] env[63088]: DEBUG nova.compute.manager [None req-407352de-05ec-47cb-9bee-a7de5aaee2c3 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: a39840c1-3d60-478b-987c-2519551963cf] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1012.111384] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-407352de-05ec-47cb-9bee-a7de5aaee2c3 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: a39840c1-3d60-478b-987c-2519551963cf] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1012.112317] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3911c12-5c43-432e-88b0-332c5ea1114c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.122579] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-407352de-05ec-47cb-9bee-a7de5aaee2c3 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: a39840c1-3d60-478b-987c-2519551963cf] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1012.126330] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-723d893e-13d9-45c5-9268-260502e86adf {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.135198] env[63088]: DEBUG oslo_vmware.api [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285434, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.136767] env[63088]: DEBUG oslo_vmware.api [None req-407352de-05ec-47cb-9bee-a7de5aaee2c3 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Waiting for the task: (returnval){ [ 1012.136767] env[63088]: value = "task-1285435" [ 1012.136767] env[63088]: _type = "Task" [ 1012.136767] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.147586] env[63088]: DEBUG oslo_vmware.api [None req-407352de-05ec-47cb-9bee-a7de5aaee2c3 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285435, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.374412] env[63088]: DEBUG nova.network.neutron [None req-5fefe597-a5a6-4e93-89ab-253f7983dd2f tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: 13388466-4adb-4d56-9fc1-e3f5dc516077] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1012.426983] env[63088]: DEBUG nova.network.neutron [None req-5fefe597-a5a6-4e93-89ab-253f7983dd2f tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: 13388466-4adb-4d56-9fc1-e3f5dc516077] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1012.505636] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4043eead-2556-4461-8694-6718416fd3ce tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.756s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1012.509454] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f2b28811-7644-45df-ad39-abad0e2f8308 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.900s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1012.509698] env[63088]: DEBUG nova.objects.instance [None req-f2b28811-7644-45df-ad39-abad0e2f8308 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Lazy-loading 'resources' on Instance uuid d6fef3ef-fec8-4929-b9b8-5e63306aa51d {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1012.530993] env[63088]: INFO nova.scheduler.client.report [None req-4043eead-2556-4461-8694-6718416fd3ce tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Deleted allocations for instance 69645f50-46fc-4c15-9b39-1721f7636e31 [ 1012.538700] env[63088]: DEBUG nova.network.neutron [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: b3b1a32f-9cc1-4810-9984-fd24c40c81cd] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1012.634417] env[63088]: DEBUG oslo_vmware.api [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285434, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.688063} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.634709] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] 2b87a9f1-5f10-43c2-8bc9-6d560ec88015/2b87a9f1-5f10-43c2-8bc9-6d560ec88015.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1012.634934] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Extending root virtual disk to 1048576 {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1012.635487] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9c92d7e7-f08b-4170-a1cd-d94099e2eb05 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.643622] env[63088]: DEBUG oslo_vmware.api [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1012.643622] env[63088]: value = "task-1285436" [ 1012.643622] env[63088]: _type = "Task" [ 1012.643622] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.647046] env[63088]: DEBUG oslo_vmware.api [None req-407352de-05ec-47cb-9bee-a7de5aaee2c3 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285435, 'name': PowerOffVM_Task} progress is 100%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.655648] env[63088]: DEBUG oslo_vmware.api [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285436, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.712483] env[63088]: DEBUG nova.network.neutron [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: b3b1a32f-9cc1-4810-9984-fd24c40c81cd] Updating instance_info_cache with network_info: [{"id": "36ea0e72-8934-47fa-8982-6b888b47766e", "address": "fa:16:3e:30:96:a6", "network": {"id": "307b966a-d9e1-40fd-9313-1ad94c734308", "bridge": "br-int", "label": "tempest-ServersTestJSON-1907249197-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "03e20d606e654362acbe2b36fe499ae3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae18b41f-e73c-44f1-83dd-467c080944f4", "external-id": "nsx-vlan-transportzone-653", "segmentation_id": 653, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap36ea0e72-89", "ovs_interfaceid": "36ea0e72-8934-47fa-8982-6b888b47766e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1012.929858] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5fefe597-a5a6-4e93-89ab-253f7983dd2f tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Releasing lock "refresh_cache-13388466-4adb-4d56-9fc1-e3f5dc516077" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1012.929858] env[63088]: DEBUG nova.compute.manager [None req-5fefe597-a5a6-4e93-89ab-253f7983dd2f tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: 13388466-4adb-4d56-9fc1-e3f5dc516077] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1012.930045] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-5fefe597-a5a6-4e93-89ab-253f7983dd2f tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: 13388466-4adb-4d56-9fc1-e3f5dc516077] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1012.930902] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-692b9edd-48c7-44e4-beb8-bc04e870f3ba {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.939784] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fefe597-a5a6-4e93-89ab-253f7983dd2f tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: 13388466-4adb-4d56-9fc1-e3f5dc516077] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1012.940089] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-95519f1e-a65d-46d9-a001-2081aed18f26 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.946857] env[63088]: DEBUG oslo_vmware.api [None req-5fefe597-a5a6-4e93-89ab-253f7983dd2f tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Waiting for the task: (returnval){ [ 1012.946857] env[63088]: value = "task-1285437" [ 1012.946857] env[63088]: _type = "Task" [ 1012.946857] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.958295] env[63088]: DEBUG oslo_vmware.api [None req-5fefe597-a5a6-4e93-89ab-253f7983dd2f tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Task: {'id': task-1285437, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.039830] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4043eead-2556-4461-8694-6718416fd3ce tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lock "69645f50-46fc-4c15-9b39-1721f7636e31" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.518s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1013.131961] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13982d33-b764-4998-be12-749bdd65d96e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.143180] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3abdab74-a358-4cd2-b0d6-0403e7a13c9e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.178122] env[63088]: DEBUG oslo_vmware.api [None req-407352de-05ec-47cb-9bee-a7de5aaee2c3 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285435, 'name': PowerOffVM_Task, 'duration_secs': 0.631409} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.179248] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-030914b0-ccf7-4383-a299-8bd588bd8f3d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.181889] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-407352de-05ec-47cb-9bee-a7de5aaee2c3 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: a39840c1-3d60-478b-987c-2519551963cf] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1013.182099] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-407352de-05ec-47cb-9bee-a7de5aaee2c3 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: a39840c1-3d60-478b-987c-2519551963cf] Unregistering the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1013.184861] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0089a00a-dc8c-49d6-ac8c-41c184858917 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.186287] env[63088]: DEBUG oslo_vmware.api [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285436, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.26815} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.186867] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Extended root virtual disk {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1013.188196] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da88091a-c950-4d87-9b4f-eca6f795b676 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.193988] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36689629-11aa-4235-be26-686384553770 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.214977] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Reconfiguring VM instance instance-00000063 to attach disk [datastore1] 2b87a9f1-5f10-43c2-8bc9-6d560ec88015/2b87a9f1-5f10-43c2-8bc9-6d560ec88015.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1013.215921] env[63088]: DEBUG oslo_concurrency.lockutils [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Releasing lock "refresh_cache-b3b1a32f-9cc1-4810-9984-fd24c40c81cd" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1013.216257] env[63088]: DEBUG nova.compute.manager [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: b3b1a32f-9cc1-4810-9984-fd24c40c81cd] Instance network_info: |[{"id": "36ea0e72-8934-47fa-8982-6b888b47766e", "address": "fa:16:3e:30:96:a6", "network": {"id": "307b966a-d9e1-40fd-9313-1ad94c734308", "bridge": "br-int", "label": "tempest-ServersTestJSON-1907249197-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "03e20d606e654362acbe2b36fe499ae3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae18b41f-e73c-44f1-83dd-467c080944f4", "external-id": "nsx-vlan-transportzone-653", "segmentation_id": 653, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap36ea0e72-89", "ovs_interfaceid": "36ea0e72-8934-47fa-8982-6b888b47766e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1013.216516] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bdfd7019-113e-4903-8fbd-241ccdc1523a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.238254] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: b3b1a32f-9cc1-4810-9984-fd24c40c81cd] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:30:96:a6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ae18b41f-e73c-44f1-83dd-467c080944f4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '36ea0e72-8934-47fa-8982-6b888b47766e', 'vif_model': 'vmxnet3'}] {{(pid=63088) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1013.245904] env[63088]: DEBUG oslo.service.loopingcall [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1013.246399] env[63088]: DEBUG nova.compute.provider_tree [None req-f2b28811-7644-45df-ad39-abad0e2f8308 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1013.247593] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b3b1a32f-9cc1-4810-9984-fd24c40c81cd] Creating VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1013.248598] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-053bda4b-7d82-4ffd-8c3c-25fbca8b38a3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.269931] env[63088]: DEBUG oslo_vmware.api [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1013.269931] env[63088]: value = "task-1285439" [ 1013.269931] env[63088]: _type = "Task" [ 1013.269931] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.274548] env[63088]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1013.274548] env[63088]: value = "task-1285440" [ 1013.274548] env[63088]: _type = "Task" [ 1013.274548] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.281643] env[63088]: DEBUG oslo_vmware.api [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285439, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.286904] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-407352de-05ec-47cb-9bee-a7de5aaee2c3 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: a39840c1-3d60-478b-987c-2519551963cf] Unregistered the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1013.287092] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-407352de-05ec-47cb-9bee-a7de5aaee2c3 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: a39840c1-3d60-478b-987c-2519551963cf] Deleting contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1013.287374] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-407352de-05ec-47cb-9bee-a7de5aaee2c3 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Deleting the datastore file [datastore1] a39840c1-3d60-478b-987c-2519551963cf {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1013.290570] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ed2cfd9a-7b9e-47db-91ce-ebeff4688436 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.292750] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285440, 'name': CreateVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.299655] env[63088]: DEBUG oslo_vmware.api [None req-407352de-05ec-47cb-9bee-a7de5aaee2c3 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Waiting for the task: (returnval){ [ 1013.299655] env[63088]: value = "task-1285441" [ 1013.299655] env[63088]: _type = "Task" [ 1013.299655] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.309106] env[63088]: DEBUG oslo_vmware.api [None req-407352de-05ec-47cb-9bee-a7de5aaee2c3 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285441, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.394921] env[63088]: DEBUG oslo_concurrency.lockutils [None req-25edd863-faa4-408b-9367-f717bec4bc40 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquiring lock "32418785-8823-4a5e-90a9-e205752530b7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1013.395191] env[63088]: DEBUG oslo_concurrency.lockutils [None req-25edd863-faa4-408b-9367-f717bec4bc40 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "32418785-8823-4a5e-90a9-e205752530b7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.395533] env[63088]: DEBUG oslo_concurrency.lockutils [None req-25edd863-faa4-408b-9367-f717bec4bc40 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquiring lock "32418785-8823-4a5e-90a9-e205752530b7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1013.395797] env[63088]: DEBUG oslo_concurrency.lockutils [None req-25edd863-faa4-408b-9367-f717bec4bc40 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "32418785-8823-4a5e-90a9-e205752530b7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.398501] env[63088]: DEBUG oslo_concurrency.lockutils [None req-25edd863-faa4-408b-9367-f717bec4bc40 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "32418785-8823-4a5e-90a9-e205752530b7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1013.398866] env[63088]: INFO nova.compute.manager [None req-25edd863-faa4-408b-9367-f717bec4bc40 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 32418785-8823-4a5e-90a9-e205752530b7] Terminating instance [ 1013.401537] env[63088]: DEBUG nova.compute.manager [None req-25edd863-faa4-408b-9367-f717bec4bc40 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 32418785-8823-4a5e-90a9-e205752530b7] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1013.402164] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-25edd863-faa4-408b-9367-f717bec4bc40 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 32418785-8823-4a5e-90a9-e205752530b7] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1013.403073] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65d8eec4-b1fc-4bc7-b37e-c7eb8d5462cc {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.408695] env[63088]: DEBUG nova.compute.manager [req-d1391ffd-3c6a-4c3e-9964-f686851ee14d req-59a8778e-c6e7-458b-b025-87d1ed9058ee service nova] [instance: b3b1a32f-9cc1-4810-9984-fd24c40c81cd] Received event network-changed-36ea0e72-8934-47fa-8982-6b888b47766e {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1013.408695] env[63088]: DEBUG nova.compute.manager [req-d1391ffd-3c6a-4c3e-9964-f686851ee14d req-59a8778e-c6e7-458b-b025-87d1ed9058ee service nova] [instance: b3b1a32f-9cc1-4810-9984-fd24c40c81cd] Refreshing instance network info cache due to event network-changed-36ea0e72-8934-47fa-8982-6b888b47766e. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1013.408814] env[63088]: DEBUG oslo_concurrency.lockutils [req-d1391ffd-3c6a-4c3e-9964-f686851ee14d req-59a8778e-c6e7-458b-b025-87d1ed9058ee service nova] Acquiring lock "refresh_cache-b3b1a32f-9cc1-4810-9984-fd24c40c81cd" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1013.408920] env[63088]: DEBUG oslo_concurrency.lockutils [req-d1391ffd-3c6a-4c3e-9964-f686851ee14d req-59a8778e-c6e7-458b-b025-87d1ed9058ee service nova] Acquired lock "refresh_cache-b3b1a32f-9cc1-4810-9984-fd24c40c81cd" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1013.409103] env[63088]: DEBUG nova.network.neutron [req-d1391ffd-3c6a-4c3e-9964-f686851ee14d req-59a8778e-c6e7-458b-b025-87d1ed9058ee service nova] [instance: b3b1a32f-9cc1-4810-9984-fd24c40c81cd] Refreshing network info cache for port 36ea0e72-8934-47fa-8982-6b888b47766e {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1013.415950] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-25edd863-faa4-408b-9367-f717bec4bc40 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 32418785-8823-4a5e-90a9-e205752530b7] Unregistering the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1013.416224] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2a95596b-9307-4a9f-b653-1571366dea5c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.457726] env[63088]: DEBUG oslo_vmware.api [None req-5fefe597-a5a6-4e93-89ab-253f7983dd2f tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Task: {'id': task-1285437, 'name': PowerOffVM_Task, 'duration_secs': 0.120432} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.458121] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fefe597-a5a6-4e93-89ab-253f7983dd2f tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: 13388466-4adb-4d56-9fc1-e3f5dc516077] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1013.458236] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-5fefe597-a5a6-4e93-89ab-253f7983dd2f tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: 13388466-4adb-4d56-9fc1-e3f5dc516077] Unregistering the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1013.458475] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d20d6c5c-2ee9-4a29-a47c-286aadc4b967 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.494959] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-5fefe597-a5a6-4e93-89ab-253f7983dd2f tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: 13388466-4adb-4d56-9fc1-e3f5dc516077] Unregistered the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1013.495983] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-5fefe597-a5a6-4e93-89ab-253f7983dd2f tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: 13388466-4adb-4d56-9fc1-e3f5dc516077] Deleting contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1013.495983] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-5fefe597-a5a6-4e93-89ab-253f7983dd2f tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Deleting the datastore file [datastore1] 13388466-4adb-4d56-9fc1-e3f5dc516077 {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1013.497495] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0d06138e-a9d6-4afe-ae58-0493626c20de {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.500192] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-25edd863-faa4-408b-9367-f717bec4bc40 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 32418785-8823-4a5e-90a9-e205752530b7] Unregistered the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1013.500379] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-25edd863-faa4-408b-9367-f717bec4bc40 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 32418785-8823-4a5e-90a9-e205752530b7] Deleting contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1013.500614] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-25edd863-faa4-408b-9367-f717bec4bc40 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Deleting the datastore file [datastore1] 32418785-8823-4a5e-90a9-e205752530b7 {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1013.500960] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5a475967-be91-4aa4-918b-6ce2b9024d1c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.509342] env[63088]: DEBUG oslo_vmware.api [None req-25edd863-faa4-408b-9367-f717bec4bc40 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 1013.509342] env[63088]: value = "task-1285445" [ 1013.509342] env[63088]: _type = "Task" [ 1013.509342] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.510938] env[63088]: DEBUG oslo_vmware.api [None req-5fefe597-a5a6-4e93-89ab-253f7983dd2f tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Waiting for the task: (returnval){ [ 1013.510938] env[63088]: value = "task-1285444" [ 1013.510938] env[63088]: _type = "Task" [ 1013.510938] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.526361] env[63088]: DEBUG oslo_vmware.api [None req-5fefe597-a5a6-4e93-89ab-253f7983dd2f tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Task: {'id': task-1285444, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.529730] env[63088]: DEBUG oslo_vmware.api [None req-25edd863-faa4-408b-9367-f717bec4bc40 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285445, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.750323] env[63088]: DEBUG nova.scheduler.client.report [None req-f2b28811-7644-45df-ad39-abad0e2f8308 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1013.781615] env[63088]: DEBUG oslo_vmware.api [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285439, 'name': ReconfigVM_Task, 'duration_secs': 0.377659} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.782154] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Reconfigured VM instance instance-00000063 to attach disk [datastore1] 2b87a9f1-5f10-43c2-8bc9-6d560ec88015/2b87a9f1-5f10-43c2-8bc9-6d560ec88015.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1013.783015] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-aad2925d-6eae-40c4-b036-29f70984e6dc {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.787382] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285440, 'name': CreateVM_Task, 'duration_secs': 0.374397} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.787854] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b3b1a32f-9cc1-4810-9984-fd24c40c81cd] Created VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1013.788556] env[63088]: DEBUG oslo_concurrency.lockutils [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1013.788735] env[63088]: DEBUG oslo_concurrency.lockutils [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1013.789041] env[63088]: DEBUG oslo_concurrency.lockutils [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1013.789285] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-00c577c5-4312-4b83-a790-3151516a3470 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.793769] env[63088]: DEBUG oslo_vmware.api [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1013.793769] env[63088]: value = "task-1285446" [ 1013.793769] env[63088]: _type = "Task" [ 1013.793769] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.794672] env[63088]: DEBUG oslo_vmware.api [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Waiting for the task: (returnval){ [ 1013.794672] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]522a69a1-3fb6-8f29-1138-73950ecaa2dd" [ 1013.794672] env[63088]: _type = "Task" [ 1013.794672] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.806690] env[63088]: DEBUG oslo_vmware.api [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285446, 'name': Rename_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.810269] env[63088]: DEBUG oslo_vmware.api [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]522a69a1-3fb6-8f29-1138-73950ecaa2dd, 'name': SearchDatastore_Task, 'duration_secs': 0.012382} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.810535] env[63088]: DEBUG oslo_concurrency.lockutils [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1013.810790] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: b3b1a32f-9cc1-4810-9984-fd24c40c81cd] Processing image 1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1013.811052] env[63088]: DEBUG oslo_concurrency.lockutils [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1013.811227] env[63088]: DEBUG oslo_concurrency.lockutils [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1013.811414] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1013.811714] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8c97199a-a744-4902-b91e-c4b484cb4622 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.815919] env[63088]: DEBUG oslo_vmware.api [None req-407352de-05ec-47cb-9bee-a7de5aaee2c3 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Task: {'id': task-1285441, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.18768} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.816493] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-407352de-05ec-47cb-9bee-a7de5aaee2c3 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1013.816707] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-407352de-05ec-47cb-9bee-a7de5aaee2c3 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: a39840c1-3d60-478b-987c-2519551963cf] Deleted contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1013.816907] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-407352de-05ec-47cb-9bee-a7de5aaee2c3 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: a39840c1-3d60-478b-987c-2519551963cf] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1013.817097] env[63088]: INFO nova.compute.manager [None req-407352de-05ec-47cb-9bee-a7de5aaee2c3 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] [instance: a39840c1-3d60-478b-987c-2519551963cf] Took 1.71 seconds to destroy the instance on the hypervisor. [ 1013.817328] env[63088]: DEBUG oslo.service.loopingcall [None req-407352de-05ec-47cb-9bee-a7de5aaee2c3 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1013.817513] env[63088]: DEBUG nova.compute.manager [-] [instance: a39840c1-3d60-478b-987c-2519551963cf] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1013.817606] env[63088]: DEBUG nova.network.neutron [-] [instance: a39840c1-3d60-478b-987c-2519551963cf] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1013.822659] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1013.822853] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1013.823592] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e2d80bf8-04d0-4317-9de0-224e9693cde1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.828978] env[63088]: DEBUG oslo_vmware.api [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Waiting for the task: (returnval){ [ 1013.828978] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52e200b4-7481-a401-fdc4-477fc0b9a094" [ 1013.828978] env[63088]: _type = "Task" [ 1013.828978] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.836423] env[63088]: DEBUG oslo_vmware.api [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52e200b4-7481-a401-fdc4-477fc0b9a094, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.023990] env[63088]: DEBUG oslo_vmware.api [None req-25edd863-faa4-408b-9367-f717bec4bc40 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285445, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.173829} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.027400] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-25edd863-faa4-408b-9367-f717bec4bc40 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1014.027737] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-25edd863-faa4-408b-9367-f717bec4bc40 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 32418785-8823-4a5e-90a9-e205752530b7] Deleted contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1014.027877] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-25edd863-faa4-408b-9367-f717bec4bc40 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 32418785-8823-4a5e-90a9-e205752530b7] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1014.027998] env[63088]: INFO nova.compute.manager [None req-25edd863-faa4-408b-9367-f717bec4bc40 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: 32418785-8823-4a5e-90a9-e205752530b7] Took 0.63 seconds to destroy the instance on the hypervisor. [ 1014.028277] env[63088]: DEBUG oslo.service.loopingcall [None req-25edd863-faa4-408b-9367-f717bec4bc40 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1014.028844] env[63088]: DEBUG oslo_vmware.api [None req-5fefe597-a5a6-4e93-89ab-253f7983dd2f tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Task: {'id': task-1285444, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.10723} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.031164] env[63088]: DEBUG nova.compute.manager [-] [instance: 32418785-8823-4a5e-90a9-e205752530b7] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1014.031258] env[63088]: DEBUG nova.network.neutron [-] [instance: 32418785-8823-4a5e-90a9-e205752530b7] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1014.032866] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-5fefe597-a5a6-4e93-89ab-253f7983dd2f tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1014.033131] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-5fefe597-a5a6-4e93-89ab-253f7983dd2f tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: 13388466-4adb-4d56-9fc1-e3f5dc516077] Deleted contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1014.033364] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-5fefe597-a5a6-4e93-89ab-253f7983dd2f tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: 13388466-4adb-4d56-9fc1-e3f5dc516077] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1014.033573] env[63088]: INFO nova.compute.manager [None req-5fefe597-a5a6-4e93-89ab-253f7983dd2f tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] [instance: 13388466-4adb-4d56-9fc1-e3f5dc516077] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1014.033807] env[63088]: DEBUG oslo.service.loopingcall [None req-5fefe597-a5a6-4e93-89ab-253f7983dd2f tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1014.037023] env[63088]: DEBUG nova.compute.manager [-] [instance: 13388466-4adb-4d56-9fc1-e3f5dc516077] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1014.037023] env[63088]: DEBUG nova.network.neutron [-] [instance: 13388466-4adb-4d56-9fc1-e3f5dc516077] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1014.056079] env[63088]: DEBUG nova.network.neutron [-] [instance: 13388466-4adb-4d56-9fc1-e3f5dc516077] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1014.245558] env[63088]: DEBUG nova.network.neutron [req-d1391ffd-3c6a-4c3e-9964-f686851ee14d req-59a8778e-c6e7-458b-b025-87d1ed9058ee service nova] [instance: b3b1a32f-9cc1-4810-9984-fd24c40c81cd] Updated VIF entry in instance network info cache for port 36ea0e72-8934-47fa-8982-6b888b47766e. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1014.245961] env[63088]: DEBUG nova.network.neutron [req-d1391ffd-3c6a-4c3e-9964-f686851ee14d req-59a8778e-c6e7-458b-b025-87d1ed9058ee service nova] [instance: b3b1a32f-9cc1-4810-9984-fd24c40c81cd] Updating instance_info_cache with network_info: [{"id": "36ea0e72-8934-47fa-8982-6b888b47766e", "address": "fa:16:3e:30:96:a6", "network": {"id": "307b966a-d9e1-40fd-9313-1ad94c734308", "bridge": "br-int", "label": "tempest-ServersTestJSON-1907249197-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "03e20d606e654362acbe2b36fe499ae3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae18b41f-e73c-44f1-83dd-467c080944f4", "external-id": "nsx-vlan-transportzone-653", "segmentation_id": 653, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap36ea0e72-89", "ovs_interfaceid": "36ea0e72-8934-47fa-8982-6b888b47766e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1014.254170] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f2b28811-7644-45df-ad39-abad0e2f8308 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.745s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1014.256283] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 3.431s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1014.275776] env[63088]: INFO nova.scheduler.client.report [None req-f2b28811-7644-45df-ad39-abad0e2f8308 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Deleted allocations for instance d6fef3ef-fec8-4929-b9b8-5e63306aa51d [ 1014.309032] env[63088]: DEBUG oslo_vmware.api [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285446, 'name': Rename_Task, 'duration_secs': 0.166913} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.309032] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1014.309032] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-764ee275-38e2-400f-84a0-c6993beec5ec {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.316443] env[63088]: DEBUG oslo_vmware.api [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1014.316443] env[63088]: value = "task-1285447" [ 1014.316443] env[63088]: _type = "Task" [ 1014.316443] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.325281] env[63088]: DEBUG oslo_vmware.api [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285447, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.340204] env[63088]: DEBUG oslo_vmware.api [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52e200b4-7481-a401-fdc4-477fc0b9a094, 'name': SearchDatastore_Task, 'duration_secs': 0.008957} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.341110] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bc7e6169-df21-412b-b5dc-967fa09d6d6c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.348170] env[63088]: DEBUG oslo_vmware.api [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Waiting for the task: (returnval){ [ 1014.348170] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]521176a7-30c2-142e-8f57-3ef9d92be4b2" [ 1014.348170] env[63088]: _type = "Task" [ 1014.348170] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.358676] env[63088]: DEBUG oslo_vmware.api [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]521176a7-30c2-142e-8f57-3ef9d92be4b2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.378048] env[63088]: DEBUG nova.compute.manager [req-271c2cb0-cad9-4356-afd2-6d06036fb555 req-ebf350b5-566b-4b02-b6e5-1822c0c8fe65 service nova] [instance: 32418785-8823-4a5e-90a9-e205752530b7] Received event network-vif-deleted-1e6d9172-35f9-4d1f-97ee-409ada1e1cd0 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1014.378221] env[63088]: INFO nova.compute.manager [req-271c2cb0-cad9-4356-afd2-6d06036fb555 req-ebf350b5-566b-4b02-b6e5-1822c0c8fe65 service nova] [instance: 32418785-8823-4a5e-90a9-e205752530b7] Neutron deleted interface 1e6d9172-35f9-4d1f-97ee-409ada1e1cd0; detaching it from the instance and deleting it from the info cache [ 1014.378360] env[63088]: DEBUG nova.network.neutron [req-271c2cb0-cad9-4356-afd2-6d06036fb555 req-ebf350b5-566b-4b02-b6e5-1822c0c8fe65 service nova] [instance: 32418785-8823-4a5e-90a9-e205752530b7] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1014.559014] env[63088]: DEBUG nova.network.neutron [-] [instance: 13388466-4adb-4d56-9fc1-e3f5dc516077] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1014.560360] env[63088]: DEBUG nova.network.neutron [-] [instance: a39840c1-3d60-478b-987c-2519551963cf] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1014.589790] env[63088]: DEBUG nova.compute.manager [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Stashing vm_state: active {{(pid=63088) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1014.748842] env[63088]: DEBUG oslo_concurrency.lockutils [req-d1391ffd-3c6a-4c3e-9964-f686851ee14d req-59a8778e-c6e7-458b-b025-87d1ed9058ee service nova] Releasing lock "refresh_cache-b3b1a32f-9cc1-4810-9984-fd24c40c81cd" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1014.783555] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f2b28811-7644-45df-ad39-abad0e2f8308 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Lock "d6fef3ef-fec8-4929-b9b8-5e63306aa51d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.342s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1014.827586] env[63088]: DEBUG oslo_vmware.api [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285447, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.854099] env[63088]: DEBUG nova.network.neutron [-] [instance: 32418785-8823-4a5e-90a9-e205752530b7] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1014.859961] env[63088]: DEBUG oslo_vmware.api [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]521176a7-30c2-142e-8f57-3ef9d92be4b2, 'name': SearchDatastore_Task, 'duration_secs': 0.014085} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.859961] env[63088]: DEBUG oslo_concurrency.lockutils [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1014.860239] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] b3b1a32f-9cc1-4810-9984-fd24c40c81cd/b3b1a32f-9cc1-4810-9984-fd24c40c81cd.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1014.860343] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bc681c32-f729-4656-903d-6fba49c7130f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.867651] env[63088]: DEBUG oslo_vmware.api [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Waiting for the task: (returnval){ [ 1014.867651] env[63088]: value = "task-1285448" [ 1014.867651] env[63088]: _type = "Task" [ 1014.867651] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.876609] env[63088]: DEBUG oslo_vmware.api [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285448, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.880281] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-15e705d6-b0d2-4f1b-bf6c-3ac3b53111c9 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.889469] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-323dfd1e-38d2-43c1-af84-1196f854bfce {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.918658] env[63088]: DEBUG nova.compute.manager [req-271c2cb0-cad9-4356-afd2-6d06036fb555 req-ebf350b5-566b-4b02-b6e5-1822c0c8fe65 service nova] [instance: 32418785-8823-4a5e-90a9-e205752530b7] Detach interface failed, port_id=1e6d9172-35f9-4d1f-97ee-409ada1e1cd0, reason: Instance 32418785-8823-4a5e-90a9-e205752530b7 could not be found. {{(pid=63088) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1015.062949] env[63088]: INFO nova.compute.manager [-] [instance: 13388466-4adb-4d56-9fc1-e3f5dc516077] Took 1.03 seconds to deallocate network for instance. [ 1015.063471] env[63088]: INFO nova.compute.manager [-] [instance: a39840c1-3d60-478b-987c-2519551963cf] Took 1.25 seconds to deallocate network for instance. [ 1015.112835] env[63088]: DEBUG oslo_concurrency.lockutils [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1015.266443] env[63088]: INFO nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Updating resource usage from migration ca2aec3e-211c-45fc-8604-4efb5c74ebbe [ 1015.286685] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance f582da15-dfc3-45e1-a995-9dd0c9533869 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1015.286936] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance 71e37d8e-a454-46c4-a3cc-3d5671a32beb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1015.287139] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance 13388466-4adb-4d56-9fc1-e3f5dc516077 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1015.287277] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance a39840c1-3d60-478b-987c-2519551963cf actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1015.287397] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance 32418785-8823-4a5e-90a9-e205752530b7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1015.287513] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance 2b87a9f1-5f10-43c2-8bc9-6d560ec88015 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1015.287629] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance b3b1a32f-9cc1-4810-9984-fd24c40c81cd actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1015.287840] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Migration ca2aec3e-211c-45fc-8604-4efb5c74ebbe is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1015.287996] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance 7b6aadb7-e34b-42b7-b69f-370434f5b665 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1015.288372] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Total usable vcpus: 48, total allocated vcpus: 8 {{(pid=63088) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1015.288577] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2048MB phys_disk=200GB used_disk=8GB total_vcpus=48 used_vcpus=8 pci_stats=[] {{(pid=63088) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1015.331656] env[63088]: DEBUG oslo_vmware.api [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285447, 'name': PowerOnVM_Task, 'duration_secs': 0.891902} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.332027] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1015.332246] env[63088]: INFO nova.compute.manager [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Took 8.59 seconds to spawn the instance on the hypervisor. [ 1015.332435] env[63088]: DEBUG nova.compute.manager [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1015.333467] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e998d82-d64e-4589-a089-40e8235c905b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.360049] env[63088]: INFO nova.compute.manager [-] [instance: 32418785-8823-4a5e-90a9-e205752530b7] Took 1.33 seconds to deallocate network for instance. [ 1015.379085] env[63088]: DEBUG oslo_vmware.api [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285448, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.426115] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64ed14b1-7a77-415f-aff2-f25a203f2271 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.431622] env[63088]: DEBUG nova.compute.manager [req-fdaeb1f9-b7a1-4972-8b1d-3b0b273385a9 req-2001361c-3c35-4cb0-b235-054f161a8e4c service nova] [instance: a39840c1-3d60-478b-987c-2519551963cf] Received event network-vif-deleted-f19d02ac-6732-44b5-b454-df5777588a7e {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1015.435573] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85125bd1-51f1-4a02-8df0-9d2e210bd055 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.468648] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5240aea1-06b1-42b6-bc37-45dfe0edba49 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.476403] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c94974a8-3da9-4ae2-9217-1782ce9b132c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.489821] env[63088]: DEBUG nova.compute.provider_tree [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1015.573939] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5fefe597-a5a6-4e93-89ab-253f7983dd2f tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1015.574724] env[63088]: DEBUG oslo_concurrency.lockutils [None req-407352de-05ec-47cb-9bee-a7de5aaee2c3 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1015.644712] env[63088]: DEBUG oslo_concurrency.lockutils [None req-75dd3299-361f-4088-bed2-ad27116ddfe6 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquiring lock "f582da15-dfc3-45e1-a995-9dd0c9533869" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1015.644997] env[63088]: DEBUG oslo_concurrency.lockutils [None req-75dd3299-361f-4088-bed2-ad27116ddfe6 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Lock "f582da15-dfc3-45e1-a995-9dd0c9533869" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1015.645259] env[63088]: DEBUG oslo_concurrency.lockutils [None req-75dd3299-361f-4088-bed2-ad27116ddfe6 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquiring lock "f582da15-dfc3-45e1-a995-9dd0c9533869-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1015.645422] env[63088]: DEBUG oslo_concurrency.lockutils [None req-75dd3299-361f-4088-bed2-ad27116ddfe6 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Lock "f582da15-dfc3-45e1-a995-9dd0c9533869-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1015.645596] env[63088]: DEBUG oslo_concurrency.lockutils [None req-75dd3299-361f-4088-bed2-ad27116ddfe6 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Lock "f582da15-dfc3-45e1-a995-9dd0c9533869-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1015.647939] env[63088]: INFO nova.compute.manager [None req-75dd3299-361f-4088-bed2-ad27116ddfe6 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Terminating instance [ 1015.650022] env[63088]: DEBUG nova.compute.manager [None req-75dd3299-361f-4088-bed2-ad27116ddfe6 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1015.650195] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-75dd3299-361f-4088-bed2-ad27116ddfe6 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1015.652057] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49939163-53e8-46a8-936a-b56ed119a295 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.660535] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-75dd3299-361f-4088-bed2-ad27116ddfe6 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1015.660797] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-05c1dce7-e72f-4da5-8285-e2fd7eda4a61 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.667277] env[63088]: DEBUG oslo_vmware.api [None req-75dd3299-361f-4088-bed2-ad27116ddfe6 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Waiting for the task: (returnval){ [ 1015.667277] env[63088]: value = "task-1285449" [ 1015.667277] env[63088]: _type = "Task" [ 1015.667277] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.676956] env[63088]: DEBUG oslo_vmware.api [None req-75dd3299-361f-4088-bed2-ad27116ddfe6 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285449, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.860437] env[63088]: INFO nova.compute.manager [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Took 15.30 seconds to build instance. [ 1015.873753] env[63088]: DEBUG oslo_concurrency.lockutils [None req-25edd863-faa4-408b-9367-f717bec4bc40 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1015.880059] env[63088]: DEBUG oslo_vmware.api [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285448, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.631351} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.880451] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] b3b1a32f-9cc1-4810-9984-fd24c40c81cd/b3b1a32f-9cc1-4810-9984-fd24c40c81cd.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1015.880584] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: b3b1a32f-9cc1-4810-9984-fd24c40c81cd] Extending root virtual disk to 1048576 {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1015.880805] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d1d6b8b7-84f1-4a47-8473-3350b9cb0c0d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.888078] env[63088]: DEBUG oslo_vmware.api [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Waiting for the task: (returnval){ [ 1015.888078] env[63088]: value = "task-1285450" [ 1015.888078] env[63088]: _type = "Task" [ 1015.888078] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.896679] env[63088]: DEBUG oslo_vmware.api [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285450, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.992971] env[63088]: DEBUG nova.scheduler.client.report [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1016.177629] env[63088]: DEBUG oslo_vmware.api [None req-75dd3299-361f-4088-bed2-ad27116ddfe6 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285449, 'name': PowerOffVM_Task, 'duration_secs': 0.212071} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.177949] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-75dd3299-361f-4088-bed2-ad27116ddfe6 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1016.178102] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-75dd3299-361f-4088-bed2-ad27116ddfe6 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Unregistering the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1016.178365] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-756be149-0954-4ddd-a263-1b04695c8391 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.246877] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-75dd3299-361f-4088-bed2-ad27116ddfe6 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Unregistered the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1016.246877] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-75dd3299-361f-4088-bed2-ad27116ddfe6 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Deleting contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1016.246877] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-75dd3299-361f-4088-bed2-ad27116ddfe6 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Deleting the datastore file [datastore1] f582da15-dfc3-45e1-a995-9dd0c9533869 {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1016.246877] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-461f5daf-5006-4ad1-aced-5af51bde4c93 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.254482] env[63088]: DEBUG oslo_vmware.api [None req-75dd3299-361f-4088-bed2-ad27116ddfe6 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Waiting for the task: (returnval){ [ 1016.254482] env[63088]: value = "task-1285452" [ 1016.254482] env[63088]: _type = "Task" [ 1016.254482] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.274344] env[63088]: DEBUG oslo_vmware.api [None req-75dd3299-361f-4088-bed2-ad27116ddfe6 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285452, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.363646] env[63088]: DEBUG oslo_concurrency.lockutils [None req-05956b8e-3e8d-40c3-953a-61cce178991a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lock "2b87a9f1-5f10-43c2-8bc9-6d560ec88015" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.813s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1016.398199] env[63088]: DEBUG oslo_vmware.api [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285450, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064908} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.398898] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: b3b1a32f-9cc1-4810-9984-fd24c40c81cd] Extended root virtual disk {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1016.399697] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8439cc82-30f1-479d-984d-c6ed61bd386f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.422564] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: b3b1a32f-9cc1-4810-9984-fd24c40c81cd] Reconfiguring VM instance instance-00000064 to attach disk [datastore1] b3b1a32f-9cc1-4810-9984-fd24c40c81cd/b3b1a32f-9cc1-4810-9984-fd24c40c81cd.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1016.422857] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-edd8afda-4a3d-43aa-9a35-1c82b9d92ec7 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.442390] env[63088]: DEBUG oslo_vmware.api [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Waiting for the task: (returnval){ [ 1016.442390] env[63088]: value = "task-1285453" [ 1016.442390] env[63088]: _type = "Task" [ 1016.442390] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.452799] env[63088]: DEBUG oslo_vmware.api [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285453, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.497310] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63088) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1016.497564] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.241s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1016.497897] env[63088]: DEBUG oslo_concurrency.lockutils [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 1.385s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1016.766118] env[63088]: DEBUG oslo_vmware.api [None req-75dd3299-361f-4088-bed2-ad27116ddfe6 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Task: {'id': task-1285452, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.331585} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.766401] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-75dd3299-361f-4088-bed2-ad27116ddfe6 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1016.766640] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-75dd3299-361f-4088-bed2-ad27116ddfe6 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Deleted contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1016.766767] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-75dd3299-361f-4088-bed2-ad27116ddfe6 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1016.766946] env[63088]: INFO nova.compute.manager [None req-75dd3299-361f-4088-bed2-ad27116ddfe6 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1016.767261] env[63088]: DEBUG oslo.service.loopingcall [None req-75dd3299-361f-4088-bed2-ad27116ddfe6 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1016.767514] env[63088]: DEBUG nova.compute.manager [-] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1016.767606] env[63088]: DEBUG nova.network.neutron [-] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1016.953931] env[63088]: DEBUG oslo_vmware.api [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285453, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.003638] env[63088]: INFO nova.compute.claims [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1017.037017] env[63088]: DEBUG nova.compute.manager [req-40645ef1-65cc-4471-ae25-0b5d96a4e4d3 req-15f06c0d-c535-4877-98b0-c65296c5c077 service nova] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Received event network-vif-deleted-81fc2dda-617e-4e4c-aa91-7e7020003995 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1017.037319] env[63088]: INFO nova.compute.manager [req-40645ef1-65cc-4471-ae25-0b5d96a4e4d3 req-15f06c0d-c535-4877-98b0-c65296c5c077 service nova] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Neutron deleted interface 81fc2dda-617e-4e4c-aa91-7e7020003995; detaching it from the instance and deleting it from the info cache [ 1017.037532] env[63088]: DEBUG nova.network.neutron [req-40645ef1-65cc-4471-ae25-0b5d96a4e4d3 req-15f06c0d-c535-4877-98b0-c65296c5c077 service nova] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1017.455226] env[63088]: DEBUG oslo_vmware.api [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285453, 'name': ReconfigVM_Task, 'duration_secs': 0.636859} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.456570] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: b3b1a32f-9cc1-4810-9984-fd24c40c81cd] Reconfigured VM instance instance-00000064 to attach disk [datastore1] b3b1a32f-9cc1-4810-9984-fd24c40c81cd/b3b1a32f-9cc1-4810-9984-fd24c40c81cd.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1017.458087] env[63088]: DEBUG nova.compute.manager [req-7418bc13-e5b5-4491-9810-fc3ef3cb0079 req-6e7ca604-3459-4b20-80a5-f8363fbf8c46 service nova] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Received event network-changed-f88388b7-bc70-4095-a792-8d155ef34534 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1017.458286] env[63088]: DEBUG nova.compute.manager [req-7418bc13-e5b5-4491-9810-fc3ef3cb0079 req-6e7ca604-3459-4b20-80a5-f8363fbf8c46 service nova] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Refreshing instance network info cache due to event network-changed-f88388b7-bc70-4095-a792-8d155ef34534. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1017.458501] env[63088]: DEBUG oslo_concurrency.lockutils [req-7418bc13-e5b5-4491-9810-fc3ef3cb0079 req-6e7ca604-3459-4b20-80a5-f8363fbf8c46 service nova] Acquiring lock "refresh_cache-2b87a9f1-5f10-43c2-8bc9-6d560ec88015" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1017.458642] env[63088]: DEBUG oslo_concurrency.lockutils [req-7418bc13-e5b5-4491-9810-fc3ef3cb0079 req-6e7ca604-3459-4b20-80a5-f8363fbf8c46 service nova] Acquired lock "refresh_cache-2b87a9f1-5f10-43c2-8bc9-6d560ec88015" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1017.458809] env[63088]: DEBUG nova.network.neutron [req-7418bc13-e5b5-4491-9810-fc3ef3cb0079 req-6e7ca604-3459-4b20-80a5-f8363fbf8c46 service nova] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Refreshing network info cache for port f88388b7-bc70-4095-a792-8d155ef34534 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1017.459854] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dd7694ec-b422-4d08-b070-bc86416e1bdd {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.468722] env[63088]: DEBUG oslo_vmware.api [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Waiting for the task: (returnval){ [ 1017.468722] env[63088]: value = "task-1285454" [ 1017.468722] env[63088]: _type = "Task" [ 1017.468722] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.478856] env[63088]: DEBUG oslo_vmware.api [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285454, 'name': Rename_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.509144] env[63088]: INFO nova.compute.resource_tracker [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Updating resource usage from migration ca2aec3e-211c-45fc-8604-4efb5c74ebbe [ 1017.511957] env[63088]: DEBUG nova.network.neutron [-] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1017.540745] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8eade5d5-c5aa-48ed-91b8-2c625ab3518d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.551852] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac9bb39e-5932-45a1-896c-62c853ed6b4b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.582793] env[63088]: DEBUG nova.compute.manager [req-40645ef1-65cc-4471-ae25-0b5d96a4e4d3 req-15f06c0d-c535-4877-98b0-c65296c5c077 service nova] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Detach interface failed, port_id=81fc2dda-617e-4e4c-aa91-7e7020003995, reason: Instance f582da15-dfc3-45e1-a995-9dd0c9533869 could not be found. {{(pid=63088) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1017.637176] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a9b1e89-55d1-444b-b0a0-35449146d7de {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.645172] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa6ebadf-c41d-4d5d-a26d-42bd45daf00e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.677410] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ae1372b-2282-4844-9b60-71a2ff96df30 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.685851] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f52dc79c-cef7-4769-8c24-7471e6a0f500 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.699480] env[63088]: DEBUG nova.compute.provider_tree [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1017.978393] env[63088]: DEBUG oslo_vmware.api [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285454, 'name': Rename_Task, 'duration_secs': 0.179815} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.978742] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: b3b1a32f-9cc1-4810-9984-fd24c40c81cd] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1017.978945] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-71ddc844-b1be-4cc7-9049-de37bf3442f1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.986774] env[63088]: DEBUG oslo_vmware.api [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Waiting for the task: (returnval){ [ 1017.986774] env[63088]: value = "task-1285455" [ 1017.986774] env[63088]: _type = "Task" [ 1017.986774] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.995128] env[63088]: DEBUG oslo_vmware.api [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285455, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.014181] env[63088]: INFO nova.compute.manager [-] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Took 1.25 seconds to deallocate network for instance. [ 1018.197818] env[63088]: DEBUG nova.network.neutron [req-7418bc13-e5b5-4491-9810-fc3ef3cb0079 req-6e7ca604-3459-4b20-80a5-f8363fbf8c46 service nova] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Updated VIF entry in instance network info cache for port f88388b7-bc70-4095-a792-8d155ef34534. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1018.198290] env[63088]: DEBUG nova.network.neutron [req-7418bc13-e5b5-4491-9810-fc3ef3cb0079 req-6e7ca604-3459-4b20-80a5-f8363fbf8c46 service nova] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Updating instance_info_cache with network_info: [{"id": "f88388b7-bc70-4095-a792-8d155ef34534", "address": "fa:16:3e:f7:d1:c7", "network": {"id": "711d3d9a-86db-4aa1-87b5-481f1812dfa4", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-59376186-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.225", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a7898ddafe0d41038a6ae4277f2c2f48", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e41070eb-3ac1-4ca9-a3d0-fd65893a97de", "external-id": "nsx-vlan-transportzone-596", "segmentation_id": 596, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf88388b7-bc", "ovs_interfaceid": "f88388b7-bc70-4095-a792-8d155ef34534", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1018.202400] env[63088]: DEBUG nova.scheduler.client.report [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1018.497259] env[63088]: DEBUG oslo_vmware.api [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285455, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.521663] env[63088]: DEBUG oslo_concurrency.lockutils [None req-75dd3299-361f-4088-bed2-ad27116ddfe6 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1018.701840] env[63088]: DEBUG oslo_concurrency.lockutils [req-7418bc13-e5b5-4491-9810-fc3ef3cb0079 req-6e7ca604-3459-4b20-80a5-f8363fbf8c46 service nova] Releasing lock "refresh_cache-2b87a9f1-5f10-43c2-8bc9-6d560ec88015" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1018.707024] env[63088]: DEBUG oslo_concurrency.lockutils [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.209s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1018.707364] env[63088]: INFO nova.compute.manager [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Migrating [ 1018.719480] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5fefe597-a5a6-4e93-89ab-253f7983dd2f tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.146s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1018.719740] env[63088]: DEBUG nova.objects.instance [None req-5fefe597-a5a6-4e93-89ab-253f7983dd2f tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Lazy-loading 'resources' on Instance uuid 13388466-4adb-4d56-9fc1-e3f5dc516077 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1018.998125] env[63088]: DEBUG oslo_vmware.api [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285455, 'name': PowerOnVM_Task, 'duration_secs': 0.817915} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.998542] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: b3b1a32f-9cc1-4810-9984-fd24c40c81cd] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1018.998649] env[63088]: INFO nova.compute.manager [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: b3b1a32f-9cc1-4810-9984-fd24c40c81cd] Took 8.00 seconds to spawn the instance on the hypervisor. [ 1018.998790] env[63088]: DEBUG nova.compute.manager [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: b3b1a32f-9cc1-4810-9984-fd24c40c81cd] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1018.999589] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7c8ea2c-994c-406e-a32a-885661f45fc1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.226080] env[63088]: DEBUG oslo_concurrency.lockutils [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquiring lock "refresh_cache-7b6aadb7-e34b-42b7-b69f-370434f5b665" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1019.226080] env[63088]: DEBUG oslo_concurrency.lockutils [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquired lock "refresh_cache-7b6aadb7-e34b-42b7-b69f-370434f5b665" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1019.226532] env[63088]: DEBUG nova.network.neutron [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1019.334764] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fada6d2a-2823-448b-892e-2be4cf15a7a3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.342731] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cfc1afe-cd85-48b2-a517-9cd4dda37b7b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.373226] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ba3952b-a7f2-4bbc-b8b5-521016b59d23 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.381294] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cba37cba-cec0-477f-8f73-bf5134c654bc {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.395371] env[63088]: DEBUG nova.compute.provider_tree [None req-5fefe597-a5a6-4e93-89ab-253f7983dd2f tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1019.520129] env[63088]: INFO nova.compute.manager [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: b3b1a32f-9cc1-4810-9984-fd24c40c81cd] Took 14.20 seconds to build instance. [ 1019.898482] env[63088]: DEBUG nova.scheduler.client.report [None req-5fefe597-a5a6-4e93-89ab-253f7983dd2f tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1019.935176] env[63088]: DEBUG nova.network.neutron [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Updating instance_info_cache with network_info: [{"id": "fa4a91d1-df9a-4789-bc5c-a8b95457cd93", "address": "fa:16:3e:7e:c2:c7", "network": {"id": "dfcbe153-bc01-4362-9247-de9f0b1c847b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-841757397-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.152", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dace8b5181b84623b08f903d12dfd31e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c68b7663-4f0e-47f0-ac7f-40c6d952f7bb", "external-id": "nsx-vlan-transportzone-696", "segmentation_id": 696, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfa4a91d1-df", "ovs_interfaceid": "fa4a91d1-df9a-4789-bc5c-a8b95457cd93", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1020.022746] env[63088]: DEBUG oslo_concurrency.lockutils [None req-517f5cc7-57ea-40df-a806-b8175a609ef3 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lock "b3b1a32f-9cc1-4810-9984-fd24c40c81cd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.712s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.150884] env[63088]: DEBUG oslo_concurrency.lockutils [None req-400cb5a1-ba24-4a60-a2ee-ee6bec0e420b tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquiring lock "b3b1a32f-9cc1-4810-9984-fd24c40c81cd" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.151364] env[63088]: DEBUG oslo_concurrency.lockutils [None req-400cb5a1-ba24-4a60-a2ee-ee6bec0e420b tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lock "b3b1a32f-9cc1-4810-9984-fd24c40c81cd" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.151696] env[63088]: DEBUG nova.compute.manager [None req-400cb5a1-ba24-4a60-a2ee-ee6bec0e420b tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: b3b1a32f-9cc1-4810-9984-fd24c40c81cd] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1020.153142] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-232353a4-b47f-43f7-b00b-96f288e4cc5d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.162959] env[63088]: DEBUG nova.compute.manager [None req-400cb5a1-ba24-4a60-a2ee-ee6bec0e420b tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: b3b1a32f-9cc1-4810-9984-fd24c40c81cd] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=63088) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1020.163800] env[63088]: DEBUG nova.objects.instance [None req-400cb5a1-ba24-4a60-a2ee-ee6bec0e420b tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lazy-loading 'flavor' on Instance uuid b3b1a32f-9cc1-4810-9984-fd24c40c81cd {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1020.405526] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5fefe597-a5a6-4e93-89ab-253f7983dd2f tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.686s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.409068] env[63088]: DEBUG oslo_concurrency.lockutils [None req-407352de-05ec-47cb-9bee-a7de5aaee2c3 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.833s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.409360] env[63088]: DEBUG nova.objects.instance [None req-407352de-05ec-47cb-9bee-a7de5aaee2c3 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Lazy-loading 'resources' on Instance uuid a39840c1-3d60-478b-987c-2519551963cf {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1020.424903] env[63088]: INFO nova.scheduler.client.report [None req-5fefe597-a5a6-4e93-89ab-253f7983dd2f tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Deleted allocations for instance 13388466-4adb-4d56-9fc1-e3f5dc516077 [ 1020.437447] env[63088]: DEBUG oslo_concurrency.lockutils [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Releasing lock "refresh_cache-7b6aadb7-e34b-42b7-b69f-370434f5b665" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1020.669994] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-400cb5a1-ba24-4a60-a2ee-ee6bec0e420b tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: b3b1a32f-9cc1-4810-9984-fd24c40c81cd] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1020.670349] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-08195e5d-d062-42c5-acfc-911498d72f49 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.678909] env[63088]: DEBUG oslo_vmware.api [None req-400cb5a1-ba24-4a60-a2ee-ee6bec0e420b tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Waiting for the task: (returnval){ [ 1020.678909] env[63088]: value = "task-1285456" [ 1020.678909] env[63088]: _type = "Task" [ 1020.678909] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.687966] env[63088]: DEBUG oslo_vmware.api [None req-400cb5a1-ba24-4a60-a2ee-ee6bec0e420b tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285456, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.944862] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5fefe597-a5a6-4e93-89ab-253f7983dd2f tempest-ServerShowV247Test-229270952 tempest-ServerShowV247Test-229270952-project-member] Lock "13388466-4adb-4d56-9fc1-e3f5dc516077" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.099s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1021.030994] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35f6d345-17c7-4105-960c-9a84c4df359c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.039542] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d59d9b5-bcc7-4b29-841f-56bc8e0c1a63 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.071357] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8a3a7b9-e641-49ce-b130-5e5ad5daa9b3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.079167] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd0a8238-0946-4bfc-ad5d-80c9a28d855f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.092665] env[63088]: DEBUG nova.compute.provider_tree [None req-407352de-05ec-47cb-9bee-a7de5aaee2c3 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1021.190211] env[63088]: DEBUG oslo_vmware.api [None req-400cb5a1-ba24-4a60-a2ee-ee6bec0e420b tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285456, 'name': PowerOffVM_Task, 'duration_secs': 0.198941} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.190489] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-400cb5a1-ba24-4a60-a2ee-ee6bec0e420b tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: b3b1a32f-9cc1-4810-9984-fd24c40c81cd] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1021.190681] env[63088]: DEBUG nova.compute.manager [None req-400cb5a1-ba24-4a60-a2ee-ee6bec0e420b tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: b3b1a32f-9cc1-4810-9984-fd24c40c81cd] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1021.191555] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fd02ae6-e2f9-4c47-9a31-d67fd48544d6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.595319] env[63088]: DEBUG nova.scheduler.client.report [None req-407352de-05ec-47cb-9bee-a7de5aaee2c3 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1021.704039] env[63088]: DEBUG oslo_concurrency.lockutils [None req-400cb5a1-ba24-4a60-a2ee-ee6bec0e420b tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lock "b3b1a32f-9cc1-4810-9984-fd24c40c81cd" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.553s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1021.954646] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd2280b8-0ee8-4f22-9ba6-b339273ec429 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.975181] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Updating instance '7b6aadb7-e34b-42b7-b69f-370434f5b665' progress to 0 {{(pid=63088) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1022.100215] env[63088]: DEBUG oslo_concurrency.lockutils [None req-407352de-05ec-47cb-9bee-a7de5aaee2c3 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.692s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1022.102583] env[63088]: DEBUG oslo_concurrency.lockutils [None req-25edd863-faa4-408b-9367-f717bec4bc40 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.229s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1022.103260] env[63088]: DEBUG nova.objects.instance [None req-25edd863-faa4-408b-9367-f717bec4bc40 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lazy-loading 'resources' on Instance uuid 32418785-8823-4a5e-90a9-e205752530b7 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1022.125835] env[63088]: INFO nova.scheduler.client.report [None req-407352de-05ec-47cb-9bee-a7de5aaee2c3 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Deleted allocations for instance a39840c1-3d60-478b-987c-2519551963cf [ 1022.483085] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1022.483363] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cfe51ac6-4639-4cd7-a561-5f8a76c2445b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.491907] env[63088]: DEBUG oslo_vmware.api [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for the task: (returnval){ [ 1022.491907] env[63088]: value = "task-1285457" [ 1022.491907] env[63088]: _type = "Task" [ 1022.491907] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.501202] env[63088]: DEBUG oslo_vmware.api [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285457, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.634009] env[63088]: DEBUG oslo_concurrency.lockutils [None req-407352de-05ec-47cb-9bee-a7de5aaee2c3 tempest-ServerDiskConfigTestJSON-1586955515 tempest-ServerDiskConfigTestJSON-1586955515-project-member] Lock "a39840c1-3d60-478b-987c-2519551963cf" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.529s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1022.703913] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ce83029-8753-4dd0-bf8f-4aa2aa2892cb {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.713038] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13156365-6fdf-40f6-ae6b-ff413ecf1066 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.744815] env[63088]: DEBUG oslo_concurrency.lockutils [None req-8f7f2e77-a82e-4ff1-abea-9dfd99ac9951 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquiring lock "b3b1a32f-9cc1-4810-9984-fd24c40c81cd" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1022.745080] env[63088]: DEBUG oslo_concurrency.lockutils [None req-8f7f2e77-a82e-4ff1-abea-9dfd99ac9951 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lock "b3b1a32f-9cc1-4810-9984-fd24c40c81cd" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1022.745289] env[63088]: DEBUG oslo_concurrency.lockutils [None req-8f7f2e77-a82e-4ff1-abea-9dfd99ac9951 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquiring lock "b3b1a32f-9cc1-4810-9984-fd24c40c81cd-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1022.745571] env[63088]: DEBUG oslo_concurrency.lockutils [None req-8f7f2e77-a82e-4ff1-abea-9dfd99ac9951 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lock "b3b1a32f-9cc1-4810-9984-fd24c40c81cd-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1022.745640] env[63088]: DEBUG oslo_concurrency.lockutils [None req-8f7f2e77-a82e-4ff1-abea-9dfd99ac9951 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lock "b3b1a32f-9cc1-4810-9984-fd24c40c81cd-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1022.747714] env[63088]: INFO nova.compute.manager [None req-8f7f2e77-a82e-4ff1-abea-9dfd99ac9951 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: b3b1a32f-9cc1-4810-9984-fd24c40c81cd] Terminating instance [ 1022.749466] env[63088]: DEBUG nova.compute.manager [None req-8f7f2e77-a82e-4ff1-abea-9dfd99ac9951 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: b3b1a32f-9cc1-4810-9984-fd24c40c81cd] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1022.749666] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-8f7f2e77-a82e-4ff1-abea-9dfd99ac9951 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: b3b1a32f-9cc1-4810-9984-fd24c40c81cd] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1022.750784] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b17c6e60-38a1-463d-a427-60098007fa8f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.753993] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5080d736-13a1-4344-ad39-542561bf28e9 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.764663] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-603d4c49-b1d7-419a-b71c-da77372bf1d1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.768698] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-8f7f2e77-a82e-4ff1-abea-9dfd99ac9951 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: b3b1a32f-9cc1-4810-9984-fd24c40c81cd] Unregistering the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1022.768698] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-61cf2c7d-7ba1-490f-92bd-d86f39eb8828 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.780502] env[63088]: DEBUG nova.compute.provider_tree [None req-25edd863-faa4-408b-9367-f717bec4bc40 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1022.849652] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-8f7f2e77-a82e-4ff1-abea-9dfd99ac9951 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: b3b1a32f-9cc1-4810-9984-fd24c40c81cd] Unregistered the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1022.849918] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-8f7f2e77-a82e-4ff1-abea-9dfd99ac9951 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: b3b1a32f-9cc1-4810-9984-fd24c40c81cd] Deleting contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1022.850170] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-8f7f2e77-a82e-4ff1-abea-9dfd99ac9951 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Deleting the datastore file [datastore1] b3b1a32f-9cc1-4810-9984-fd24c40c81cd {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1022.850497] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0c83173e-3480-4168-b7f4-de2bbd2bac01 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.858186] env[63088]: DEBUG oslo_vmware.api [None req-8f7f2e77-a82e-4ff1-abea-9dfd99ac9951 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Waiting for the task: (returnval){ [ 1022.858186] env[63088]: value = "task-1285459" [ 1022.858186] env[63088]: _type = "Task" [ 1022.858186] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.866890] env[63088]: DEBUG oslo_vmware.api [None req-8f7f2e77-a82e-4ff1-abea-9dfd99ac9951 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285459, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.004977] env[63088]: DEBUG oslo_vmware.api [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285457, 'name': PowerOffVM_Task, 'duration_secs': 0.176042} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.005322] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1023.005528] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Updating instance '7b6aadb7-e34b-42b7-b69f-370434f5b665' progress to 17 {{(pid=63088) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1023.284105] env[63088]: DEBUG nova.scheduler.client.report [None req-25edd863-faa4-408b-9367-f717bec4bc40 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1023.370491] env[63088]: DEBUG oslo_vmware.api [None req-8f7f2e77-a82e-4ff1-abea-9dfd99ac9951 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Task: {'id': task-1285459, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.167609} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.370787] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-8f7f2e77-a82e-4ff1-abea-9dfd99ac9951 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1023.371027] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-8f7f2e77-a82e-4ff1-abea-9dfd99ac9951 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: b3b1a32f-9cc1-4810-9984-fd24c40c81cd] Deleted contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1023.371249] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-8f7f2e77-a82e-4ff1-abea-9dfd99ac9951 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: b3b1a32f-9cc1-4810-9984-fd24c40c81cd] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1023.371448] env[63088]: INFO nova.compute.manager [None req-8f7f2e77-a82e-4ff1-abea-9dfd99ac9951 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] [instance: b3b1a32f-9cc1-4810-9984-fd24c40c81cd] Took 0.62 seconds to destroy the instance on the hypervisor. [ 1023.371744] env[63088]: DEBUG oslo.service.loopingcall [None req-8f7f2e77-a82e-4ff1-abea-9dfd99ac9951 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1023.371969] env[63088]: DEBUG nova.compute.manager [-] [instance: b3b1a32f-9cc1-4810-9984-fd24c40c81cd] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1023.372109] env[63088]: DEBUG nova.network.neutron [-] [instance: b3b1a32f-9cc1-4810-9984-fd24c40c81cd] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1023.512820] env[63088]: DEBUG nova.virt.hardware [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:33Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1023.513064] env[63088]: DEBUG nova.virt.hardware [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1023.513243] env[63088]: DEBUG nova.virt.hardware [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1023.513428] env[63088]: DEBUG nova.virt.hardware [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1023.513578] env[63088]: DEBUG nova.virt.hardware [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1023.513730] env[63088]: DEBUG nova.virt.hardware [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1023.513935] env[63088]: DEBUG nova.virt.hardware [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1023.514114] env[63088]: DEBUG nova.virt.hardware [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1023.514286] env[63088]: DEBUG nova.virt.hardware [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1023.514453] env[63088]: DEBUG nova.virt.hardware [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1023.514627] env[63088]: DEBUG nova.virt.hardware [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1023.519793] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1fca8cdd-cca0-49df-8ead-06f5af5c3915 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.537944] env[63088]: DEBUG oslo_vmware.api [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for the task: (returnval){ [ 1023.537944] env[63088]: value = "task-1285460" [ 1023.537944] env[63088]: _type = "Task" [ 1023.537944] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.553197] env[63088]: DEBUG oslo_vmware.api [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285460, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.646230] env[63088]: DEBUG nova.compute.manager [req-13805ffd-288f-42b3-9a19-484238330d06 req-52a5f422-b55a-4264-b90e-0baab6a65e5d service nova] [instance: b3b1a32f-9cc1-4810-9984-fd24c40c81cd] Received event network-vif-deleted-36ea0e72-8934-47fa-8982-6b888b47766e {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1023.646373] env[63088]: INFO nova.compute.manager [req-13805ffd-288f-42b3-9a19-484238330d06 req-52a5f422-b55a-4264-b90e-0baab6a65e5d service nova] [instance: b3b1a32f-9cc1-4810-9984-fd24c40c81cd] Neutron deleted interface 36ea0e72-8934-47fa-8982-6b888b47766e; detaching it from the instance and deleting it from the info cache [ 1023.646551] env[63088]: DEBUG nova.network.neutron [req-13805ffd-288f-42b3-9a19-484238330d06 req-52a5f422-b55a-4264-b90e-0baab6a65e5d service nova] [instance: b3b1a32f-9cc1-4810-9984-fd24c40c81cd] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1023.789368] env[63088]: DEBUG oslo_concurrency.lockutils [None req-25edd863-faa4-408b-9367-f717bec4bc40 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.686s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1023.791477] env[63088]: DEBUG oslo_concurrency.lockutils [None req-75dd3299-361f-4088-bed2-ad27116ddfe6 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.270s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1023.791727] env[63088]: DEBUG nova.objects.instance [None req-75dd3299-361f-4088-bed2-ad27116ddfe6 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Lazy-loading 'resources' on Instance uuid f582da15-dfc3-45e1-a995-9dd0c9533869 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1023.811405] env[63088]: INFO nova.scheduler.client.report [None req-25edd863-faa4-408b-9367-f717bec4bc40 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Deleted allocations for instance 32418785-8823-4a5e-90a9-e205752530b7 [ 1024.049529] env[63088]: DEBUG oslo_vmware.api [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285460, 'name': ReconfigVM_Task, 'duration_secs': 0.457157} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.049856] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Updating instance '7b6aadb7-e34b-42b7-b69f-370434f5b665' progress to 33 {{(pid=63088) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1024.125509] env[63088]: DEBUG nova.network.neutron [-] [instance: b3b1a32f-9cc1-4810-9984-fd24c40c81cd] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1024.150706] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-97c6dcca-4da4-48d3-a606-6c73239fbea4 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.164487] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-365f4c1f-9137-4e44-8d10-2c2fbe351d8f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.194835] env[63088]: DEBUG nova.compute.manager [req-13805ffd-288f-42b3-9a19-484238330d06 req-52a5f422-b55a-4264-b90e-0baab6a65e5d service nova] [instance: b3b1a32f-9cc1-4810-9984-fd24c40c81cd] Detach interface failed, port_id=36ea0e72-8934-47fa-8982-6b888b47766e, reason: Instance b3b1a32f-9cc1-4810-9984-fd24c40c81cd could not be found. {{(pid=63088) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1024.321252] env[63088]: DEBUG oslo_concurrency.lockutils [None req-25edd863-faa4-408b-9367-f717bec4bc40 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "32418785-8823-4a5e-90a9-e205752530b7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.926s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1024.381076] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d64058b0-cb08-44f1-8561-f927fdbea8c0 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.389095] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-944ee5a3-5e84-406b-bda9-c6c050224a2e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.421029] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1830ed67-c8c5-4ed2-ab1f-f24b3cbe44f2 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.429334] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76d1adb1-8238-4740-8d96-58ade1d185f1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.443362] env[63088]: DEBUG nova.compute.provider_tree [None req-75dd3299-361f-4088-bed2-ad27116ddfe6 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1024.556771] env[63088]: DEBUG nova.virt.hardware [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1024.557059] env[63088]: DEBUG nova.virt.hardware [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1024.557230] env[63088]: DEBUG nova.virt.hardware [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1024.557419] env[63088]: DEBUG nova.virt.hardware [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1024.557570] env[63088]: DEBUG nova.virt.hardware [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1024.557757] env[63088]: DEBUG nova.virt.hardware [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1024.557956] env[63088]: DEBUG nova.virt.hardware [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1024.558140] env[63088]: DEBUG nova.virt.hardware [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1024.558312] env[63088]: DEBUG nova.virt.hardware [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1024.558478] env[63088]: DEBUG nova.virt.hardware [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1024.558753] env[63088]: DEBUG nova.virt.hardware [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1024.564585] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Reconfiguring VM instance instance-00000042 to detach disk 2000 {{(pid=63088) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1024.564585] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3eeda001-1977-432e-8ce6-ad7fbaebeacb {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.585321] env[63088]: DEBUG oslo_vmware.api [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for the task: (returnval){ [ 1024.585321] env[63088]: value = "task-1285461" [ 1024.585321] env[63088]: _type = "Task" [ 1024.585321] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.598571] env[63088]: DEBUG oslo_vmware.api [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285461, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.628309] env[63088]: INFO nova.compute.manager [-] [instance: b3b1a32f-9cc1-4810-9984-fd24c40c81cd] Took 1.26 seconds to deallocate network for instance. [ 1024.946951] env[63088]: DEBUG nova.scheduler.client.report [None req-75dd3299-361f-4088-bed2-ad27116ddfe6 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1025.096566] env[63088]: DEBUG oslo_vmware.api [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285461, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.134816] env[63088]: DEBUG oslo_concurrency.lockutils [None req-8f7f2e77-a82e-4ff1-abea-9dfd99ac9951 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1025.452650] env[63088]: DEBUG oslo_concurrency.lockutils [None req-75dd3299-361f-4088-bed2-ad27116ddfe6 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.661s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.455067] env[63088]: DEBUG oslo_concurrency.lockutils [None req-8f7f2e77-a82e-4ff1-abea-9dfd99ac9951 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.320s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.455314] env[63088]: DEBUG nova.objects.instance [None req-8f7f2e77-a82e-4ff1-abea-9dfd99ac9951 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lazy-loading 'resources' on Instance uuid b3b1a32f-9cc1-4810-9984-fd24c40c81cd {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1025.469788] env[63088]: INFO nova.scheduler.client.report [None req-75dd3299-361f-4088-bed2-ad27116ddfe6 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Deleted allocations for instance f582da15-dfc3-45e1-a995-9dd0c9533869 [ 1025.595787] env[63088]: DEBUG oslo_vmware.api [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285461, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.832897] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquiring lock "d94e57a3-743b-4bfe-b208-80f5748f8654" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1025.833145] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "d94e57a3-743b-4bfe-b208-80f5748f8654" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.979790] env[63088]: DEBUG oslo_concurrency.lockutils [None req-75dd3299-361f-4088-bed2-ad27116ddfe6 tempest-AttachInterfacesTestJSON-1262560851 tempest-AttachInterfacesTestJSON-1262560851-project-member] Lock "f582da15-dfc3-45e1-a995-9dd0c9533869" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.335s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1026.043763] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97b7297c-7b5b-4fbc-ba09-dc48f1e4b13c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.051588] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-727aa7ca-36b6-42f6-9846-e0bac59e0ecc {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.081088] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0346b5c2-e13b-45cd-a233-eb5e1c58d335 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.091138] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32a14a1f-8d89-4a57-a562-b51c8152eea2 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.099554] env[63088]: DEBUG oslo_vmware.api [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285461, 'name': ReconfigVM_Task, 'duration_secs': 1.175883} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.108408] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Reconfigured VM instance instance-00000042 to detach disk 2000 {{(pid=63088) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1026.108891] env[63088]: DEBUG nova.compute.provider_tree [None req-8f7f2e77-a82e-4ff1-abea-9dfd99ac9951 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1026.111084] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab18e2b4-0393-4ccc-8bdd-ac1187204417 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.133515] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Reconfiguring VM instance instance-00000042 to attach disk [datastore1] 7b6aadb7-e34b-42b7-b69f-370434f5b665/7b6aadb7-e34b-42b7-b69f-370434f5b665.vmdk or device None with type thin {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1026.134560] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8e7f4925-3823-4fd6-bc46-469e76370e9c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.153599] env[63088]: DEBUG oslo_vmware.api [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for the task: (returnval){ [ 1026.153599] env[63088]: value = "task-1285462" [ 1026.153599] env[63088]: _type = "Task" [ 1026.153599] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.162691] env[63088]: DEBUG oslo_vmware.api [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285462, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.335615] env[63088]: DEBUG nova.compute.manager [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1026.612249] env[63088]: DEBUG nova.scheduler.client.report [None req-8f7f2e77-a82e-4ff1-abea-9dfd99ac9951 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1026.665631] env[63088]: DEBUG oslo_vmware.api [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285462, 'name': ReconfigVM_Task, 'duration_secs': 0.284988} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.666104] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Reconfigured VM instance instance-00000042 to attach disk [datastore1] 7b6aadb7-e34b-42b7-b69f-370434f5b665/7b6aadb7-e34b-42b7-b69f-370434f5b665.vmdk or device None with type thin {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1026.666405] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Updating instance '7b6aadb7-e34b-42b7-b69f-370434f5b665' progress to 50 {{(pid=63088) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1026.856707] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1027.119027] env[63088]: DEBUG oslo_concurrency.lockutils [None req-8f7f2e77-a82e-4ff1-abea-9dfd99ac9951 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.664s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1027.121620] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.265s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1027.123159] env[63088]: INFO nova.compute.claims [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1027.143666] env[63088]: INFO nova.scheduler.client.report [None req-8f7f2e77-a82e-4ff1-abea-9dfd99ac9951 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Deleted allocations for instance b3b1a32f-9cc1-4810-9984-fd24c40c81cd [ 1027.173608] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7499819-e2d1-46c5-a59c-4c6f01379dd1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.194365] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb3bf81c-fdac-471e-ac4d-71ef3e462f3d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.213610] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Updating instance '7b6aadb7-e34b-42b7-b69f-370434f5b665' progress to 67 {{(pid=63088) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1027.651765] env[63088]: DEBUG oslo_concurrency.lockutils [None req-8f7f2e77-a82e-4ff1-abea-9dfd99ac9951 tempest-ServersTestJSON-58652973 tempest-ServersTestJSON-58652973-project-member] Lock "b3b1a32f-9cc1-4810-9984-fd24c40c81cd" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.907s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1027.766646] env[63088]: DEBUG nova.network.neutron [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Port fa4a91d1-df9a-4789-bc5c-a8b95457cd93 binding to destination host cpu-1 is already ACTIVE {{(pid=63088) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1028.217151] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45dfaee7-44af-495f-bc47-593c0a4a031a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.226018] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbc1ec4b-5353-48ba-8632-b4b08e1e713a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.256937] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a92cf15-9ae1-45de-8988-bfa87d78cc7e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.265115] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca4f882d-ea0b-4df1-8428-e8866436001c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.284419] env[63088]: DEBUG nova.compute.provider_tree [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1028.791793] env[63088]: DEBUG oslo_concurrency.lockutils [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquiring lock "7b6aadb7-e34b-42b7-b69f-370434f5b665-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1028.793920] env[63088]: DEBUG oslo_concurrency.lockutils [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lock "7b6aadb7-e34b-42b7-b69f-370434f5b665-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1028.794147] env[63088]: DEBUG oslo_concurrency.lockutils [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lock "7b6aadb7-e34b-42b7-b69f-370434f5b665-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.003s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1028.795776] env[63088]: DEBUG nova.scheduler.client.report [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1029.302619] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.181s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1029.303227] env[63088]: DEBUG nova.compute.manager [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1029.810397] env[63088]: DEBUG nova.compute.utils [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1029.812399] env[63088]: DEBUG nova.compute.manager [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1029.812399] env[63088]: DEBUG nova.network.neutron [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1029.855805] env[63088]: DEBUG oslo_concurrency.lockutils [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquiring lock "refresh_cache-7b6aadb7-e34b-42b7-b69f-370434f5b665" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1029.855909] env[63088]: DEBUG oslo_concurrency.lockutils [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquired lock "refresh_cache-7b6aadb7-e34b-42b7-b69f-370434f5b665" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1029.856169] env[63088]: DEBUG nova.network.neutron [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1029.872695] env[63088]: DEBUG nova.policy [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '139b65d9cf314b3db1d685d63bc19bba', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e48f62dee9ad4e9b94b67bd871db5c63', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 1030.153374] env[63088]: DEBUG nova.network.neutron [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Successfully created port: 024c5e7f-23a2-46df-83c1-5195ff662b68 {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1030.315130] env[63088]: DEBUG nova.compute.manager [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1030.635068] env[63088]: DEBUG nova.network.neutron [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Updating instance_info_cache with network_info: [{"id": "fa4a91d1-df9a-4789-bc5c-a8b95457cd93", "address": "fa:16:3e:7e:c2:c7", "network": {"id": "dfcbe153-bc01-4362-9247-de9f0b1c847b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-841757397-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.152", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dace8b5181b84623b08f903d12dfd31e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c68b7663-4f0e-47f0-ac7f-40c6d952f7bb", "external-id": "nsx-vlan-transportzone-696", "segmentation_id": 696, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfa4a91d1-df", "ovs_interfaceid": "fa4a91d1-df9a-4789-bc5c-a8b95457cd93", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1030.944857] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Acquiring lock "9d2e0a07-3542-4e05-9902-82a949baa5a8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1030.945311] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Lock "9d2e0a07-3542-4e05-9902-82a949baa5a8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1031.137371] env[63088]: DEBUG oslo_concurrency.lockutils [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Releasing lock "refresh_cache-7b6aadb7-e34b-42b7-b69f-370434f5b665" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1031.324915] env[63088]: DEBUG nova.compute.manager [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1031.376420] env[63088]: DEBUG nova.virt.hardware [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1031.376827] env[63088]: DEBUG nova.virt.hardware [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1031.377098] env[63088]: DEBUG nova.virt.hardware [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1031.377377] env[63088]: DEBUG nova.virt.hardware [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1031.377607] env[63088]: DEBUG nova.virt.hardware [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1031.378169] env[63088]: DEBUG nova.virt.hardware [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1031.378326] env[63088]: DEBUG nova.virt.hardware [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1031.378489] env[63088]: DEBUG nova.virt.hardware [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1031.378753] env[63088]: DEBUG nova.virt.hardware [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1031.379050] env[63088]: DEBUG nova.virt.hardware [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1031.379340] env[63088]: DEBUG nova.virt.hardware [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1031.380781] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5e5fae2-7742-4938-97c8-d29bb92ca756 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.391405] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5de3502c-cc3e-4c05-9961-f7b812ae8dc5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.448032] env[63088]: DEBUG nova.compute.manager [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] [instance: 9d2e0a07-3542-4e05-9902-82a949baa5a8] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1031.674601] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19b04f3a-10a5-4a5a-b4d8-42f223d5bc8a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.679341] env[63088]: DEBUG nova.compute.manager [req-fac2f682-8173-4c19-972e-692e596fb577 req-0ab6507e-4d61-4e6b-9fbd-4278651d68c0 service nova] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Received event network-vif-plugged-024c5e7f-23a2-46df-83c1-5195ff662b68 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1031.679558] env[63088]: DEBUG oslo_concurrency.lockutils [req-fac2f682-8173-4c19-972e-692e596fb577 req-0ab6507e-4d61-4e6b-9fbd-4278651d68c0 service nova] Acquiring lock "d94e57a3-743b-4bfe-b208-80f5748f8654-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1031.679771] env[63088]: DEBUG oslo_concurrency.lockutils [req-fac2f682-8173-4c19-972e-692e596fb577 req-0ab6507e-4d61-4e6b-9fbd-4278651d68c0 service nova] Lock "d94e57a3-743b-4bfe-b208-80f5748f8654-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1031.679940] env[63088]: DEBUG oslo_concurrency.lockutils [req-fac2f682-8173-4c19-972e-692e596fb577 req-0ab6507e-4d61-4e6b-9fbd-4278651d68c0 service nova] Lock "d94e57a3-743b-4bfe-b208-80f5748f8654-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1031.680124] env[63088]: DEBUG nova.compute.manager [req-fac2f682-8173-4c19-972e-692e596fb577 req-0ab6507e-4d61-4e6b-9fbd-4278651d68c0 service nova] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] No waiting events found dispatching network-vif-plugged-024c5e7f-23a2-46df-83c1-5195ff662b68 {{(pid=63088) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1031.680297] env[63088]: WARNING nova.compute.manager [req-fac2f682-8173-4c19-972e-692e596fb577 req-0ab6507e-4d61-4e6b-9fbd-4278651d68c0 service nova] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Received unexpected event network-vif-plugged-024c5e7f-23a2-46df-83c1-5195ff662b68 for instance with vm_state building and task_state spawning. [ 1031.706846] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df194038-bb26-419d-837f-491cae72f87f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.715421] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Updating instance '7b6aadb7-e34b-42b7-b69f-370434f5b665' progress to 83 {{(pid=63088) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1031.777055] env[63088]: DEBUG nova.network.neutron [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Successfully updated port: 024c5e7f-23a2-46df-83c1-5195ff662b68 {{(pid=63088) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1031.968842] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1031.969213] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1031.970996] env[63088]: INFO nova.compute.claims [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] [instance: 9d2e0a07-3542-4e05-9902-82a949baa5a8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1032.224331] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1032.224621] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c8ad47b3-5442-451a-92f0-1792a9c90ae8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.232628] env[63088]: DEBUG oslo_vmware.api [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for the task: (returnval){ [ 1032.232628] env[63088]: value = "task-1285463" [ 1032.232628] env[63088]: _type = "Task" [ 1032.232628] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.241371] env[63088]: DEBUG oslo_vmware.api [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285463, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.279834] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquiring lock "refresh_cache-d94e57a3-743b-4bfe-b208-80f5748f8654" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1032.279995] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquired lock "refresh_cache-d94e57a3-743b-4bfe-b208-80f5748f8654" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1032.280177] env[63088]: DEBUG nova.network.neutron [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1032.742958] env[63088]: DEBUG oslo_vmware.api [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285463, 'name': PowerOnVM_Task, 'duration_secs': 0.397995} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.743181] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1032.743378] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-fdd41b45-e9d3-4097-ada2-19b27a9b3e35 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Updating instance '7b6aadb7-e34b-42b7-b69f-370434f5b665' progress to 100 {{(pid=63088) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1032.828513] env[63088]: DEBUG nova.network.neutron [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1032.983858] env[63088]: DEBUG nova.network.neutron [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Updating instance_info_cache with network_info: [{"id": "024c5e7f-23a2-46df-83c1-5195ff662b68", "address": "fa:16:3e:76:26:9c", "network": {"id": "5b801381-6bb3-45cd-b81f-92b5d757e080", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1481457494-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e48f62dee9ad4e9b94b67bd871db5c63", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "051f343d-ac4f-4070-a26d-467603122c81", "external-id": "nsx-vlan-transportzone-277", "segmentation_id": 277, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap024c5e7f-23", "ovs_interfaceid": "024c5e7f-23a2-46df-83c1-5195ff662b68", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1033.061498] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9c26899-9072-4f8c-ac29-9dd5df6d1fb8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.069720] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1a3d9e5-4578-4582-a127-762c177bf571 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.099425] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ae1e3ee-17fc-40da-849c-5a50bc20b1f4 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.106716] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6020d4fb-79b5-4a36-a0b8-d8d9c1f68848 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.119703] env[63088]: DEBUG nova.compute.provider_tree [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1033.486154] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Releasing lock "refresh_cache-d94e57a3-743b-4bfe-b208-80f5748f8654" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1033.486578] env[63088]: DEBUG nova.compute.manager [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Instance network_info: |[{"id": "024c5e7f-23a2-46df-83c1-5195ff662b68", "address": "fa:16:3e:76:26:9c", "network": {"id": "5b801381-6bb3-45cd-b81f-92b5d757e080", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1481457494-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e48f62dee9ad4e9b94b67bd871db5c63", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "051f343d-ac4f-4070-a26d-467603122c81", "external-id": "nsx-vlan-transportzone-277", "segmentation_id": 277, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap024c5e7f-23", "ovs_interfaceid": "024c5e7f-23a2-46df-83c1-5195ff662b68", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1033.487072] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:76:26:9c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '051f343d-ac4f-4070-a26d-467603122c81', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '024c5e7f-23a2-46df-83c1-5195ff662b68', 'vif_model': 'vmxnet3'}] {{(pid=63088) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1033.494720] env[63088]: DEBUG oslo.service.loopingcall [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1033.494965] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Creating VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1033.495243] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f5e74aef-fd4b-4548-8853-9ee3bb65b2ae {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.516106] env[63088]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1033.516106] env[63088]: value = "task-1285464" [ 1033.516106] env[63088]: _type = "Task" [ 1033.516106] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.525063] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285464, 'name': CreateVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.623559] env[63088]: DEBUG nova.scheduler.client.report [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1033.704132] env[63088]: DEBUG nova.compute.manager [req-eabd4b96-f8e2-4cf6-84a4-64a3927a864a req-3b4874c2-94e3-4f90-8fc1-a2938e461829 service nova] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Received event network-changed-024c5e7f-23a2-46df-83c1-5195ff662b68 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1033.704359] env[63088]: DEBUG nova.compute.manager [req-eabd4b96-f8e2-4cf6-84a4-64a3927a864a req-3b4874c2-94e3-4f90-8fc1-a2938e461829 service nova] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Refreshing instance network info cache due to event network-changed-024c5e7f-23a2-46df-83c1-5195ff662b68. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1033.704677] env[63088]: DEBUG oslo_concurrency.lockutils [req-eabd4b96-f8e2-4cf6-84a4-64a3927a864a req-3b4874c2-94e3-4f90-8fc1-a2938e461829 service nova] Acquiring lock "refresh_cache-d94e57a3-743b-4bfe-b208-80f5748f8654" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1033.704828] env[63088]: DEBUG oslo_concurrency.lockutils [req-eabd4b96-f8e2-4cf6-84a4-64a3927a864a req-3b4874c2-94e3-4f90-8fc1-a2938e461829 service nova] Acquired lock "refresh_cache-d94e57a3-743b-4bfe-b208-80f5748f8654" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1033.705055] env[63088]: DEBUG nova.network.neutron [req-eabd4b96-f8e2-4cf6-84a4-64a3927a864a req-3b4874c2-94e3-4f90-8fc1-a2938e461829 service nova] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Refreshing network info cache for port 024c5e7f-23a2-46df-83c1-5195ff662b68 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1034.027307] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285464, 'name': CreateVM_Task, 'duration_secs': 0.338131} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.027632] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Created VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1034.028186] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1034.028359] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1034.028678] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1034.028960] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0edd5e88-3186-426d-80a1-19694b602dfb {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.033936] env[63088]: DEBUG oslo_vmware.api [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 1034.033936] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]526be2bf-ac79-53f7-8518-0b13409c0daa" [ 1034.033936] env[63088]: _type = "Task" [ 1034.033936] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.041582] env[63088]: DEBUG oslo_vmware.api [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]526be2bf-ac79-53f7-8518-0b13409c0daa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.129031] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.160s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1034.129536] env[63088]: DEBUG nova.compute.manager [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] [instance: 9d2e0a07-3542-4e05-9902-82a949baa5a8] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1034.391674] env[63088]: DEBUG nova.network.neutron [req-eabd4b96-f8e2-4cf6-84a4-64a3927a864a req-3b4874c2-94e3-4f90-8fc1-a2938e461829 service nova] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Updated VIF entry in instance network info cache for port 024c5e7f-23a2-46df-83c1-5195ff662b68. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1034.392182] env[63088]: DEBUG nova.network.neutron [req-eabd4b96-f8e2-4cf6-84a4-64a3927a864a req-3b4874c2-94e3-4f90-8fc1-a2938e461829 service nova] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Updating instance_info_cache with network_info: [{"id": "024c5e7f-23a2-46df-83c1-5195ff662b68", "address": "fa:16:3e:76:26:9c", "network": {"id": "5b801381-6bb3-45cd-b81f-92b5d757e080", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1481457494-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e48f62dee9ad4e9b94b67bd871db5c63", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "051f343d-ac4f-4070-a26d-467603122c81", "external-id": "nsx-vlan-transportzone-277", "segmentation_id": 277, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap024c5e7f-23", "ovs_interfaceid": "024c5e7f-23a2-46df-83c1-5195ff662b68", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1034.547454] env[63088]: DEBUG oslo_vmware.api [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]526be2bf-ac79-53f7-8518-0b13409c0daa, 'name': SearchDatastore_Task, 'duration_secs': 0.011077} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.547873] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1034.548226] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Processing image 1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1034.548568] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1034.548800] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1034.549252] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1034.549564] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0b619db8-fbbf-4e3f-82f7-e3922e13d738 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.557870] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1034.558067] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1034.558766] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1ee5a1e0-8ee4-47b1-85e8-ff71f96d0de1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.563766] env[63088]: DEBUG oslo_vmware.api [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 1034.563766] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]5243a55f-e53b-3e18-2303-cf41683fd1c4" [ 1034.563766] env[63088]: _type = "Task" [ 1034.563766] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.571799] env[63088]: DEBUG oslo_vmware.api [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]5243a55f-e53b-3e18-2303-cf41683fd1c4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.634277] env[63088]: DEBUG nova.compute.utils [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1034.635728] env[63088]: DEBUG nova.compute.manager [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] [instance: 9d2e0a07-3542-4e05-9902-82a949baa5a8] Not allocating networking since 'none' was specified. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 1034.894845] env[63088]: DEBUG oslo_concurrency.lockutils [req-eabd4b96-f8e2-4cf6-84a4-64a3927a864a req-3b4874c2-94e3-4f90-8fc1-a2938e461829 service nova] Releasing lock "refresh_cache-d94e57a3-743b-4bfe-b208-80f5748f8654" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1035.074619] env[63088]: DEBUG oslo_vmware.api [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]5243a55f-e53b-3e18-2303-cf41683fd1c4, 'name': SearchDatastore_Task, 'duration_secs': 0.00917} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.075436] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-10999526-1a48-46c7-825c-ccc12c35e345 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.081037] env[63088]: DEBUG oslo_vmware.api [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 1035.081037] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52f60f8c-258e-2138-8c2c-8b3c11cb33f7" [ 1035.081037] env[63088]: _type = "Task" [ 1035.081037] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.088880] env[63088]: DEBUG oslo_vmware.api [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52f60f8c-258e-2138-8c2c-8b3c11cb33f7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.138907] env[63088]: DEBUG nova.compute.manager [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] [instance: 9d2e0a07-3542-4e05-9902-82a949baa5a8] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1035.411117] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b5a40319-4a97-4a1f-9c98-12f22f896c8e tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquiring lock "7b6aadb7-e34b-42b7-b69f-370434f5b665" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.413172] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b5a40319-4a97-4a1f-9c98-12f22f896c8e tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lock "7b6aadb7-e34b-42b7-b69f-370434f5b665" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.002s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.413385] env[63088]: DEBUG nova.compute.manager [None req-b5a40319-4a97-4a1f-9c98-12f22f896c8e tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Going to confirm migration 4 {{(pid=63088) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1035.592020] env[63088]: DEBUG oslo_vmware.api [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52f60f8c-258e-2138-8c2c-8b3c11cb33f7, 'name': SearchDatastore_Task, 'duration_secs': 0.01011} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.592312] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1035.592577] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] d94e57a3-743b-4bfe-b208-80f5748f8654/d94e57a3-743b-4bfe-b208-80f5748f8654.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1035.592829] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e72d556c-0059-41d6-84fb-d7467b0e3dc0 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.601473] env[63088]: DEBUG oslo_vmware.api [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 1035.601473] env[63088]: value = "task-1285465" [ 1035.601473] env[63088]: _type = "Task" [ 1035.601473] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.608986] env[63088]: DEBUG oslo_vmware.api [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285465, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.990021] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b5a40319-4a97-4a1f-9c98-12f22f896c8e tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquiring lock "refresh_cache-7b6aadb7-e34b-42b7-b69f-370434f5b665" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1035.990259] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b5a40319-4a97-4a1f-9c98-12f22f896c8e tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquired lock "refresh_cache-7b6aadb7-e34b-42b7-b69f-370434f5b665" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1035.990476] env[63088]: DEBUG nova.network.neutron [None req-b5a40319-4a97-4a1f-9c98-12f22f896c8e tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1035.990674] env[63088]: DEBUG nova.objects.instance [None req-b5a40319-4a97-4a1f-9c98-12f22f896c8e tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lazy-loading 'info_cache' on Instance uuid 7b6aadb7-e34b-42b7-b69f-370434f5b665 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1036.112555] env[63088]: DEBUG oslo_vmware.api [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285465, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.148487] env[63088]: DEBUG nova.compute.manager [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] [instance: 9d2e0a07-3542-4e05-9902-82a949baa5a8] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1036.173304] env[63088]: DEBUG nova.virt.hardware [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1036.173581] env[63088]: DEBUG nova.virt.hardware [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1036.173742] env[63088]: DEBUG nova.virt.hardware [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1036.173929] env[63088]: DEBUG nova.virt.hardware [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1036.174771] env[63088]: DEBUG nova.virt.hardware [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1036.174771] env[63088]: DEBUG nova.virt.hardware [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1036.174771] env[63088]: DEBUG nova.virt.hardware [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1036.174771] env[63088]: DEBUG nova.virt.hardware [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1036.174948] env[63088]: DEBUG nova.virt.hardware [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1036.174989] env[63088]: DEBUG nova.virt.hardware [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1036.175177] env[63088]: DEBUG nova.virt.hardware [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1036.176098] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4635320b-0a8b-4376-8180-419cde7a6893 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.184532] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1812ac6a-9c00-48a9-8aea-6151f8f01783 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.198362] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] [instance: 9d2e0a07-3542-4e05-9902-82a949baa5a8] Instance VIF info [] {{(pid=63088) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1036.204202] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Creating folder: Project (33e556ef604048fb98684827b2abf95c). Parent ref: group-v275816. {{(pid=63088) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1036.204492] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4030ecfe-bfef-4556-9b8e-2652e19f1f85 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.215036] env[63088]: INFO nova.virt.vmwareapi.vm_util [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Created folder: Project (33e556ef604048fb98684827b2abf95c) in parent group-v275816. [ 1036.215228] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Creating folder: Instances. Parent ref: group-v275952. {{(pid=63088) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1036.215450] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6338b232-bb49-4ea4-a956-b1f76d2eb933 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.224835] env[63088]: INFO nova.virt.vmwareapi.vm_util [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Created folder: Instances in parent group-v275952. [ 1036.225084] env[63088]: DEBUG oslo.service.loopingcall [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1036.225290] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9d2e0a07-3542-4e05-9902-82a949baa5a8] Creating VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1036.225491] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-10933483-8abf-4b84-a759-41a0d8481b61 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.241262] env[63088]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1036.241262] env[63088]: value = "task-1285468" [ 1036.241262] env[63088]: _type = "Task" [ 1036.241262] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.248647] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285468, 'name': CreateVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.612722] env[63088]: DEBUG oslo_vmware.api [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285465, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.526024} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.612992] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] d94e57a3-743b-4bfe-b208-80f5748f8654/d94e57a3-743b-4bfe-b208-80f5748f8654.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1036.613232] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Extending root virtual disk to 1048576 {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1036.613481] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5d8f34a9-6060-4bf8-9250-ba7031bbf6f7 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.619518] env[63088]: DEBUG oslo_vmware.api [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 1036.619518] env[63088]: value = "task-1285469" [ 1036.619518] env[63088]: _type = "Task" [ 1036.619518] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.626734] env[63088]: DEBUG oslo_vmware.api [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285469, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.750970] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285468, 'name': CreateVM_Task, 'duration_secs': 0.305189} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.751208] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9d2e0a07-3542-4e05-9902-82a949baa5a8] Created VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1036.751672] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1036.751840] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1036.752397] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1036.752474] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f9f568f6-1564-4446-a488-e280516ef310 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.757160] env[63088]: DEBUG oslo_vmware.api [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Waiting for the task: (returnval){ [ 1036.757160] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52ddd1b3-24e8-58ae-4e69-dc7b3e659ea2" [ 1036.757160] env[63088]: _type = "Task" [ 1036.757160] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.766371] env[63088]: DEBUG oslo_vmware.api [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52ddd1b3-24e8-58ae-4e69-dc7b3e659ea2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.129687] env[63088]: DEBUG oslo_vmware.api [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285469, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063214} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.129969] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Extended root virtual disk {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1037.130726] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43a73b97-897c-45a6-ac7d-2e54d42267e2 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.154834] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Reconfiguring VM instance instance-00000065 to attach disk [datastore1] d94e57a3-743b-4bfe-b208-80f5748f8654/d94e57a3-743b-4bfe-b208-80f5748f8654.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1037.155183] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-05aebd8d-249e-4301-9195-3442473692f0 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.177503] env[63088]: DEBUG oslo_vmware.api [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 1037.177503] env[63088]: value = "task-1285470" [ 1037.177503] env[63088]: _type = "Task" [ 1037.177503] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.185723] env[63088]: DEBUG oslo_vmware.api [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285470, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.204061] env[63088]: DEBUG nova.network.neutron [None req-b5a40319-4a97-4a1f-9c98-12f22f896c8e tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Updating instance_info_cache with network_info: [{"id": "fa4a91d1-df9a-4789-bc5c-a8b95457cd93", "address": "fa:16:3e:7e:c2:c7", "network": {"id": "dfcbe153-bc01-4362-9247-de9f0b1c847b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-841757397-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.152", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dace8b5181b84623b08f903d12dfd31e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c68b7663-4f0e-47f0-ac7f-40c6d952f7bb", "external-id": "nsx-vlan-transportzone-696", "segmentation_id": 696, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfa4a91d1-df", "ovs_interfaceid": "fa4a91d1-df9a-4789-bc5c-a8b95457cd93", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1037.268768] env[63088]: DEBUG oslo_vmware.api [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52ddd1b3-24e8-58ae-4e69-dc7b3e659ea2, 'name': SearchDatastore_Task, 'duration_secs': 0.093589} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.269108] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1037.269364] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] [instance: 9d2e0a07-3542-4e05-9902-82a949baa5a8] Processing image 1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1037.269639] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1037.269846] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1037.270090] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1037.270385] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fa8be888-5044-46aa-9e23-482bae0a0b9c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.279904] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1037.280150] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1037.280967] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1c60c081-f5c6-4f50-aa93-84478d22dbbb {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.287234] env[63088]: DEBUG oslo_vmware.api [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Waiting for the task: (returnval){ [ 1037.287234] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52e618dc-c0e6-5adb-9e85-01878c1c0d8a" [ 1037.287234] env[63088]: _type = "Task" [ 1037.287234] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.295892] env[63088]: DEBUG oslo_vmware.api [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52e618dc-c0e6-5adb-9e85-01878c1c0d8a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.688490] env[63088]: DEBUG oslo_vmware.api [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285470, 'name': ReconfigVM_Task, 'duration_secs': 0.302154} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.688705] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Reconfigured VM instance instance-00000065 to attach disk [datastore1] d94e57a3-743b-4bfe-b208-80f5748f8654/d94e57a3-743b-4bfe-b208-80f5748f8654.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1037.689379] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f6887769-4efb-4c75-96a2-98b0261498f2 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.696233] env[63088]: DEBUG oslo_vmware.api [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 1037.696233] env[63088]: value = "task-1285471" [ 1037.696233] env[63088]: _type = "Task" [ 1037.696233] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.703894] env[63088]: DEBUG oslo_vmware.api [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285471, 'name': Rename_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.706503] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b5a40319-4a97-4a1f-9c98-12f22f896c8e tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Releasing lock "refresh_cache-7b6aadb7-e34b-42b7-b69f-370434f5b665" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1037.706783] env[63088]: DEBUG nova.objects.instance [None req-b5a40319-4a97-4a1f-9c98-12f22f896c8e tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lazy-loading 'migration_context' on Instance uuid 7b6aadb7-e34b-42b7-b69f-370434f5b665 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1037.799645] env[63088]: DEBUG oslo_vmware.api [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52e618dc-c0e6-5adb-9e85-01878c1c0d8a, 'name': SearchDatastore_Task, 'duration_secs': 0.015456} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.800581] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-22a9faf9-b6a5-4683-bc91-dc0409c057b8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.807223] env[63088]: DEBUG oslo_vmware.api [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Waiting for the task: (returnval){ [ 1037.807223] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52219242-2a34-d165-2643-12907f6ba0cc" [ 1037.807223] env[63088]: _type = "Task" [ 1037.807223] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.815566] env[63088]: DEBUG oslo_vmware.api [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52219242-2a34-d165-2643-12907f6ba0cc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.206969] env[63088]: DEBUG oslo_vmware.api [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285471, 'name': Rename_Task, 'duration_secs': 0.145789} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.207389] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1038.207514] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-98eb7352-db21-40c0-ad1d-d48028f67608 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.209561] env[63088]: DEBUG nova.objects.base [None req-b5a40319-4a97-4a1f-9c98-12f22f896c8e tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Object Instance<7b6aadb7-e34b-42b7-b69f-370434f5b665> lazy-loaded attributes: info_cache,migration_context {{(pid=63088) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1038.210360] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71fd6822-e38a-4ee3-992b-d7713249ebfb {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.230977] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c6f67be0-a302-47fa-bc3c-fe58f42bd2ba {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.233381] env[63088]: DEBUG oslo_vmware.api [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 1038.233381] env[63088]: value = "task-1285472" [ 1038.233381] env[63088]: _type = "Task" [ 1038.233381] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.238269] env[63088]: DEBUG oslo_vmware.api [None req-b5a40319-4a97-4a1f-9c98-12f22f896c8e tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for the task: (returnval){ [ 1038.238269] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]523e5689-ed05-f9b2-312f-6797b8450c0a" [ 1038.238269] env[63088]: _type = "Task" [ 1038.238269] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.244395] env[63088]: DEBUG oslo_vmware.api [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285472, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.249745] env[63088]: DEBUG oslo_vmware.api [None req-b5a40319-4a97-4a1f-9c98-12f22f896c8e tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]523e5689-ed05-f9b2-312f-6797b8450c0a, 'name': SearchDatastore_Task, 'duration_secs': 0.007214} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.250078] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b5a40319-4a97-4a1f-9c98-12f22f896c8e tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1038.250262] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b5a40319-4a97-4a1f-9c98-12f22f896c8e tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1038.317983] env[63088]: DEBUG oslo_vmware.api [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52219242-2a34-d165-2643-12907f6ba0cc, 'name': SearchDatastore_Task, 'duration_secs': 0.010039} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.318331] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1038.318631] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] 9d2e0a07-3542-4e05-9902-82a949baa5a8/9d2e0a07-3542-4e05-9902-82a949baa5a8.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1038.318919] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d02d5550-fb24-4796-95ba-5cdafd0e98fa {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.325540] env[63088]: DEBUG oslo_vmware.api [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Waiting for the task: (returnval){ [ 1038.325540] env[63088]: value = "task-1285473" [ 1038.325540] env[63088]: _type = "Task" [ 1038.325540] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.333256] env[63088]: DEBUG oslo_vmware.api [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Task: {'id': task-1285473, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.746283] env[63088]: DEBUG oslo_vmware.api [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285472, 'name': PowerOnVM_Task, 'duration_secs': 0.463287} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.746685] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1038.746949] env[63088]: INFO nova.compute.manager [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Took 7.42 seconds to spawn the instance on the hypervisor. [ 1038.747209] env[63088]: DEBUG nova.compute.manager [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1038.748174] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f27b0c9-222e-4f3c-9892-127dcc031bdc {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.837459] env[63088]: DEBUG oslo_vmware.api [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Task: {'id': task-1285473, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.454377} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.837738] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] 9d2e0a07-3542-4e05-9902-82a949baa5a8/9d2e0a07-3542-4e05-9902-82a949baa5a8.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1038.837956] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] [instance: 9d2e0a07-3542-4e05-9902-82a949baa5a8] Extending root virtual disk to 1048576 {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1038.838298] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b0eae3ba-7d06-47c3-9802-78e811c71564 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.841278] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f9c2ee6-aa9e-4c07-83d0-91e65ef8bed7 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.850012] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d3c0e1b-9f66-461e-ae3a-78a7a361954f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.853107] env[63088]: DEBUG oslo_vmware.api [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Waiting for the task: (returnval){ [ 1038.853107] env[63088]: value = "task-1285474" [ 1038.853107] env[63088]: _type = "Task" [ 1038.853107] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.882632] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85c18aab-902c-44e1-9bb6-79206a06c14a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.887887] env[63088]: DEBUG oslo_vmware.api [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Task: {'id': task-1285474, 'name': ExtendVirtualDisk_Task} progress is 50%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.893484] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52678dfd-2d1a-4feb-ab2c-1ca4c2134928 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.907411] env[63088]: DEBUG nova.compute.provider_tree [None req-b5a40319-4a97-4a1f-9c98-12f22f896c8e tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1039.268377] env[63088]: INFO nova.compute.manager [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Took 12.43 seconds to build instance. [ 1039.363460] env[63088]: DEBUG oslo_vmware.api [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Task: {'id': task-1285474, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.057745} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.363730] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] [instance: 9d2e0a07-3542-4e05-9902-82a949baa5a8] Extended root virtual disk {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1039.364670] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12238f4e-183a-4bab-90a7-1d3fc6a6166c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.383812] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] [instance: 9d2e0a07-3542-4e05-9902-82a949baa5a8] Reconfiguring VM instance instance-00000066 to attach disk [datastore1] 9d2e0a07-3542-4e05-9902-82a949baa5a8/9d2e0a07-3542-4e05-9902-82a949baa5a8.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1039.384058] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-df382c26-c02b-4922-bbb8-09bd8b955e07 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.402927] env[63088]: DEBUG oslo_vmware.api [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Waiting for the task: (returnval){ [ 1039.402927] env[63088]: value = "task-1285475" [ 1039.402927] env[63088]: _type = "Task" [ 1039.402927] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.410968] env[63088]: DEBUG nova.scheduler.client.report [None req-b5a40319-4a97-4a1f-9c98-12f22f896c8e tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1039.417425] env[63088]: DEBUG oslo_vmware.api [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Task: {'id': task-1285475, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.770380] env[63088]: DEBUG oslo_concurrency.lockutils [None req-bbeb5865-c4d8-4f22-b6ce-6f3ec3a27ba2 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "d94e57a3-743b-4bfe-b208-80f5748f8654" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.937s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1039.915553] env[63088]: DEBUG oslo_vmware.api [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Task: {'id': task-1285475, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.417187] env[63088]: DEBUG oslo_vmware.api [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Task: {'id': task-1285475, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.426618] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b5a40319-4a97-4a1f-9c98-12f22f896c8e tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.176s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1040.915312] env[63088]: DEBUG oslo_vmware.api [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Task: {'id': task-1285475, 'name': ReconfigVM_Task, 'duration_secs': 1.07776} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.915619] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] [instance: 9d2e0a07-3542-4e05-9902-82a949baa5a8] Reconfigured VM instance instance-00000066 to attach disk [datastore1] 9d2e0a07-3542-4e05-9902-82a949baa5a8/9d2e0a07-3542-4e05-9902-82a949baa5a8.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1040.916283] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d2d15824-154f-4e5d-a6c8-6ad50af2ccd2 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.922577] env[63088]: DEBUG oslo_vmware.api [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Waiting for the task: (returnval){ [ 1040.922577] env[63088]: value = "task-1285476" [ 1040.922577] env[63088]: _type = "Task" [ 1040.922577] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.934300] env[63088]: DEBUG oslo_vmware.api [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Task: {'id': task-1285476, 'name': Rename_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.991873] env[63088]: INFO nova.scheduler.client.report [None req-b5a40319-4a97-4a1f-9c98-12f22f896c8e tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Deleted allocation for migration ca2aec3e-211c-45fc-8604-4efb5c74ebbe [ 1041.017124] env[63088]: DEBUG nova.compute.manager [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Stashing vm_state: active {{(pid=63088) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1041.433402] env[63088]: DEBUG oslo_vmware.api [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Task: {'id': task-1285476, 'name': Rename_Task, 'duration_secs': 0.169993} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.433755] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] [instance: 9d2e0a07-3542-4e05-9902-82a949baa5a8] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1041.434014] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7c84a4ad-4713-41ff-bb2c-f8dd747453f8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.440913] env[63088]: DEBUG oslo_vmware.api [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Waiting for the task: (returnval){ [ 1041.440913] env[63088]: value = "task-1285477" [ 1041.440913] env[63088]: _type = "Task" [ 1041.440913] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.449164] env[63088]: DEBUG oslo_vmware.api [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Task: {'id': task-1285477, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.498720] env[63088]: DEBUG oslo_concurrency.lockutils [None req-b5a40319-4a97-4a1f-9c98-12f22f896c8e tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lock "7b6aadb7-e34b-42b7-b69f-370434f5b665" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.085s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1041.538358] env[63088]: DEBUG oslo_concurrency.lockutils [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1041.538666] env[63088]: DEBUG oslo_concurrency.lockutils [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1041.960700] env[63088]: DEBUG oslo_vmware.api [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Task: {'id': task-1285477, 'name': PowerOnVM_Task} progress is 78%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.043600] env[63088]: INFO nova.compute.claims [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1042.447881] env[63088]: DEBUG oslo_concurrency.lockutils [None req-22e7e5df-a740-49c0-a4c7-551def0e818d tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquiring lock "7b6aadb7-e34b-42b7-b69f-370434f5b665" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1042.448346] env[63088]: DEBUG oslo_concurrency.lockutils [None req-22e7e5df-a740-49c0-a4c7-551def0e818d tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lock "7b6aadb7-e34b-42b7-b69f-370434f5b665" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1042.448424] env[63088]: DEBUG oslo_concurrency.lockutils [None req-22e7e5df-a740-49c0-a4c7-551def0e818d tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquiring lock "7b6aadb7-e34b-42b7-b69f-370434f5b665-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1042.448545] env[63088]: DEBUG oslo_concurrency.lockutils [None req-22e7e5df-a740-49c0-a4c7-551def0e818d tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lock "7b6aadb7-e34b-42b7-b69f-370434f5b665-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1042.448718] env[63088]: DEBUG oslo_concurrency.lockutils [None req-22e7e5df-a740-49c0-a4c7-551def0e818d tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lock "7b6aadb7-e34b-42b7-b69f-370434f5b665-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1042.453233] env[63088]: DEBUG oslo_vmware.api [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Task: {'id': task-1285477, 'name': PowerOnVM_Task, 'duration_secs': 0.748777} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.453700] env[63088]: INFO nova.compute.manager [None req-22e7e5df-a740-49c0-a4c7-551def0e818d tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Terminating instance [ 1042.455014] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] [instance: 9d2e0a07-3542-4e05-9902-82a949baa5a8] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1042.455323] env[63088]: INFO nova.compute.manager [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] [instance: 9d2e0a07-3542-4e05-9902-82a949baa5a8] Took 6.31 seconds to spawn the instance on the hypervisor. [ 1042.455401] env[63088]: DEBUG nova.compute.manager [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] [instance: 9d2e0a07-3542-4e05-9902-82a949baa5a8] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1042.455969] env[63088]: DEBUG nova.compute.manager [None req-22e7e5df-a740-49c0-a4c7-551def0e818d tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1042.456178] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-22e7e5df-a740-49c0-a4c7-551def0e818d tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1042.456886] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0325aac-2ba8-494f-8f9a-27f648e67f70 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.459786] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c942566d-15ed-4899-911b-eca4e1845756 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.469675] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-22e7e5df-a740-49c0-a4c7-551def0e818d tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1042.470078] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b91709cb-3a68-43d4-87df-f5d288b9952f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.475467] env[63088]: DEBUG oslo_vmware.api [None req-22e7e5df-a740-49c0-a4c7-551def0e818d tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for the task: (returnval){ [ 1042.475467] env[63088]: value = "task-1285478" [ 1042.475467] env[63088]: _type = "Task" [ 1042.475467] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.483668] env[63088]: DEBUG oslo_vmware.api [None req-22e7e5df-a740-49c0-a4c7-551def0e818d tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285478, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.550370] env[63088]: INFO nova.compute.resource_tracker [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Updating resource usage from migration 1873436d-a9d6-4d6c-80b5-d6b5decb0c77 [ 1042.640822] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02f00b6f-5a06-45f5-b8e8-fb0a078e6516 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.648218] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aebf832e-268c-4adc-9d99-30445bca7192 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.679966] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee27f80a-96e7-4049-8ab7-ea5104299045 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.687176] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46b5cc4c-9ed0-469e-ae47-8a848104138b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.700203] env[63088]: DEBUG nova.compute.provider_tree [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1042.979935] env[63088]: INFO nova.compute.manager [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] [instance: 9d2e0a07-3542-4e05-9902-82a949baa5a8] Took 11.03 seconds to build instance. [ 1042.986583] env[63088]: DEBUG oslo_vmware.api [None req-22e7e5df-a740-49c0-a4c7-551def0e818d tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285478, 'name': PowerOffVM_Task, 'duration_secs': 0.206357} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.986872] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-22e7e5df-a740-49c0-a4c7-551def0e818d tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1042.987052] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-22e7e5df-a740-49c0-a4c7-551def0e818d tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Unregistering the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1042.987314] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-db1eb256-956b-4459-951e-97bde6f8689f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.070762] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-22e7e5df-a740-49c0-a4c7-551def0e818d tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Unregistered the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1043.071012] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-22e7e5df-a740-49c0-a4c7-551def0e818d tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Deleting contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1043.071229] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-22e7e5df-a740-49c0-a4c7-551def0e818d tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Deleting the datastore file [datastore1] 7b6aadb7-e34b-42b7-b69f-370434f5b665 {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1043.071549] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c06a3c20-7257-4cf0-a260-d42348cc8c93 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.077466] env[63088]: DEBUG oslo_vmware.api [None req-22e7e5df-a740-49c0-a4c7-551def0e818d tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for the task: (returnval){ [ 1043.077466] env[63088]: value = "task-1285480" [ 1043.077466] env[63088]: _type = "Task" [ 1043.077466] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.086282] env[63088]: DEBUG oslo_vmware.api [None req-22e7e5df-a740-49c0-a4c7-551def0e818d tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285480, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.203207] env[63088]: DEBUG nova.scheduler.client.report [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1043.482430] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5e5de303-a40c-49f3-a816-7e1803711d8d tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Lock "9d2e0a07-3542-4e05-9902-82a949baa5a8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.537s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1043.587494] env[63088]: DEBUG oslo_vmware.api [None req-22e7e5df-a740-49c0-a4c7-551def0e818d tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285480, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.142148} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.587781] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-22e7e5df-a740-49c0-a4c7-551def0e818d tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1043.588013] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-22e7e5df-a740-49c0-a4c7-551def0e818d tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Deleted contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1043.588220] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-22e7e5df-a740-49c0-a4c7-551def0e818d tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1043.588406] env[63088]: INFO nova.compute.manager [None req-22e7e5df-a740-49c0-a4c7-551def0e818d tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1043.588657] env[63088]: DEBUG oslo.service.loopingcall [None req-22e7e5df-a740-49c0-a4c7-551def0e818d tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1043.588857] env[63088]: DEBUG nova.compute.manager [-] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1043.588952] env[63088]: DEBUG nova.network.neutron [-] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1043.708042] env[63088]: DEBUG oslo_concurrency.lockutils [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.169s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1043.708291] env[63088]: INFO nova.compute.manager [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Migrating [ 1043.830301] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1043.830531] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1043.830679] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Starting heal instance info cache {{(pid=63088) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1043.830799] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Rebuilding the list of instances to heal {{(pid=63088) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 1043.854397] env[63088]: INFO nova.compute.manager [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] [instance: 9d2e0a07-3542-4e05-9902-82a949baa5a8] Rebuilding instance [ 1043.892292] env[63088]: DEBUG nova.compute.manager [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] [instance: 9d2e0a07-3542-4e05-9902-82a949baa5a8] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1043.893187] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30fe993f-7b81-448b-8ef9-622590e45c89 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.014352] env[63088]: DEBUG nova.compute.manager [req-ad701fa2-d08c-42f7-b989-0bc60a5f5262 req-b9efb34a-c778-48ba-bef6-32ee819161e5 service nova] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Received event network-vif-deleted-fa4a91d1-df9a-4789-bc5c-a8b95457cd93 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1044.014692] env[63088]: INFO nova.compute.manager [req-ad701fa2-d08c-42f7-b989-0bc60a5f5262 req-b9efb34a-c778-48ba-bef6-32ee819161e5 service nova] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Neutron deleted interface fa4a91d1-df9a-4789-bc5c-a8b95457cd93; detaching it from the instance and deleting it from the info cache [ 1044.014988] env[63088]: DEBUG nova.network.neutron [req-ad701fa2-d08c-42f7-b989-0bc60a5f5262 req-b9efb34a-c778-48ba-bef6-32ee819161e5 service nova] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1044.225321] env[63088]: DEBUG oslo_concurrency.lockutils [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquiring lock "refresh_cache-d94e57a3-743b-4bfe-b208-80f5748f8654" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1044.225525] env[63088]: DEBUG oslo_concurrency.lockutils [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquired lock "refresh_cache-d94e57a3-743b-4bfe-b208-80f5748f8654" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1044.225715] env[63088]: DEBUG nova.network.neutron [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1044.334178] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Skipping network cache update for instance because it is being deleted. {{(pid=63088) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9943}} [ 1044.360580] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Acquiring lock "refresh_cache-71e37d8e-a454-46c4-a3cc-3d5671a32beb" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1044.360732] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Acquired lock "refresh_cache-71e37d8e-a454-46c4-a3cc-3d5671a32beb" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1044.360884] env[63088]: DEBUG nova.network.neutron [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Forcefully refreshing network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1044.361048] env[63088]: DEBUG nova.objects.instance [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Lazy-loading 'info_cache' on Instance uuid 71e37d8e-a454-46c4-a3cc-3d5671a32beb {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1044.403664] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] [instance: 9d2e0a07-3542-4e05-9902-82a949baa5a8] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1044.403987] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-04e496e6-212c-49a4-bdf3-ebb039c80b30 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.411807] env[63088]: DEBUG oslo_vmware.api [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Waiting for the task: (returnval){ [ 1044.411807] env[63088]: value = "task-1285481" [ 1044.411807] env[63088]: _type = "Task" [ 1044.411807] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.420693] env[63088]: DEBUG oslo_vmware.api [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Task: {'id': task-1285481, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.490686] env[63088]: DEBUG nova.network.neutron [-] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1044.517314] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-396b2525-b4b6-4963-b966-656d2a01f8ff {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.528009] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e29c6a4a-0fa4-4977-97c5-eba2d2e07026 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.553446] env[63088]: DEBUG nova.compute.manager [req-ad701fa2-d08c-42f7-b989-0bc60a5f5262 req-b9efb34a-c778-48ba-bef6-32ee819161e5 service nova] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Detach interface failed, port_id=fa4a91d1-df9a-4789-bc5c-a8b95457cd93, reason: Instance 7b6aadb7-e34b-42b7-b69f-370434f5b665 could not be found. {{(pid=63088) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1044.921742] env[63088]: DEBUG oslo_vmware.api [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Task: {'id': task-1285481, 'name': PowerOffVM_Task, 'duration_secs': 0.184112} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.922037] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] [instance: 9d2e0a07-3542-4e05-9902-82a949baa5a8] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1044.922205] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] [instance: 9d2e0a07-3542-4e05-9902-82a949baa5a8] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1044.922983] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02fc0911-8c87-4950-9d10-fde0f7f1dd13 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.929646] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] [instance: 9d2e0a07-3542-4e05-9902-82a949baa5a8] Unregistering the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1044.929873] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-513bc8ab-0ecc-4d50-a44a-7aa605b6e715 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.955203] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] [instance: 9d2e0a07-3542-4e05-9902-82a949baa5a8] Unregistered the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1044.955472] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] [instance: 9d2e0a07-3542-4e05-9902-82a949baa5a8] Deleting contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1044.955619] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Deleting the datastore file [datastore1] 9d2e0a07-3542-4e05-9902-82a949baa5a8 {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1044.955953] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8d1085b6-df7e-47d9-a2bd-92827807393e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.962301] env[63088]: DEBUG oslo_vmware.api [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Waiting for the task: (returnval){ [ 1044.962301] env[63088]: value = "task-1285483" [ 1044.962301] env[63088]: _type = "Task" [ 1044.962301] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.970461] env[63088]: DEBUG oslo_vmware.api [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Task: {'id': task-1285483, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.993087] env[63088]: INFO nova.compute.manager [-] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Took 1.40 seconds to deallocate network for instance. [ 1045.100675] env[63088]: DEBUG nova.network.neutron [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Updating instance_info_cache with network_info: [{"id": "024c5e7f-23a2-46df-83c1-5195ff662b68", "address": "fa:16:3e:76:26:9c", "network": {"id": "5b801381-6bb3-45cd-b81f-92b5d757e080", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1481457494-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e48f62dee9ad4e9b94b67bd871db5c63", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "051f343d-ac4f-4070-a26d-467603122c81", "external-id": "nsx-vlan-transportzone-277", "segmentation_id": 277, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap024c5e7f-23", "ovs_interfaceid": "024c5e7f-23a2-46df-83c1-5195ff662b68", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1045.472199] env[63088]: DEBUG oslo_vmware.api [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Task: {'id': task-1285483, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.097178} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.472429] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1045.472593] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] [instance: 9d2e0a07-3542-4e05-9902-82a949baa5a8] Deleted contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1045.472773] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] [instance: 9d2e0a07-3542-4e05-9902-82a949baa5a8] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1045.503267] env[63088]: DEBUG oslo_concurrency.lockutils [None req-22e7e5df-a740-49c0-a4c7-551def0e818d tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1045.504109] env[63088]: DEBUG oslo_concurrency.lockutils [None req-22e7e5df-a740-49c0-a4c7-551def0e818d tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1045.504418] env[63088]: DEBUG oslo_concurrency.lockutils [None req-22e7e5df-a740-49c0-a4c7-551def0e818d tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1045.530082] env[63088]: INFO nova.scheduler.client.report [None req-22e7e5df-a740-49c0-a4c7-551def0e818d tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Deleted allocations for instance 7b6aadb7-e34b-42b7-b69f-370434f5b665 [ 1045.602924] env[63088]: DEBUG oslo_concurrency.lockutils [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Releasing lock "refresh_cache-d94e57a3-743b-4bfe-b208-80f5748f8654" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1045.746021] env[63088]: DEBUG oslo_concurrency.lockutils [None req-cf33486b-20b0-41d5-b03c-7534ec45d4ad tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Acquiring lock "71e37d8e-a454-46c4-a3cc-3d5671a32beb" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1045.746242] env[63088]: DEBUG oslo_concurrency.lockutils [None req-cf33486b-20b0-41d5-b03c-7534ec45d4ad tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Lock "71e37d8e-a454-46c4-a3cc-3d5671a32beb" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1046.037267] env[63088]: DEBUG oslo_concurrency.lockutils [None req-22e7e5df-a740-49c0-a4c7-551def0e818d tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lock "7b6aadb7-e34b-42b7-b69f-370434f5b665" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.589s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1046.085985] env[63088]: DEBUG nova.network.neutron [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Updating instance_info_cache with network_info: [{"id": "79ca8279-a3e2-4792-a4b5-57f899be5c41", "address": "fa:16:3e:53:f8:3c", "network": {"id": "cc2c0304-8906-443a-a76e-e7a32158346f", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1867758042-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.139", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b18ddddf2314d83addf550b8cb91977", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0f096917-a0cf-4add-a9d2-23ca1c723b3b", "external-id": "nsx-vlan-transportzone-894", "segmentation_id": 894, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap79ca8279-a3", "ovs_interfaceid": "79ca8279-a3e2-4792-a4b5-57f899be5c41", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1046.248862] env[63088]: INFO nova.compute.manager [None req-cf33486b-20b0-41d5-b03c-7534ec45d4ad tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Detaching volume 8a71432a-16cf-4ea7-a27f-e1899ca9f9e8 [ 1046.283435] env[63088]: INFO nova.virt.block_device [None req-cf33486b-20b0-41d5-b03c-7534ec45d4ad tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Attempting to driver detach volume 8a71432a-16cf-4ea7-a27f-e1899ca9f9e8 from mountpoint /dev/sdb [ 1046.283735] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-cf33486b-20b0-41d5-b03c-7534ec45d4ad tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Volume detach. Driver type: vmdk {{(pid=63088) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1046.283927] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-cf33486b-20b0-41d5-b03c-7534ec45d4ad tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-275932', 'volume_id': '8a71432a-16cf-4ea7-a27f-e1899ca9f9e8', 'name': 'volume-8a71432a-16cf-4ea7-a27f-e1899ca9f9e8', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '71e37d8e-a454-46c4-a3cc-3d5671a32beb', 'attached_at': '', 'detached_at': '', 'volume_id': '8a71432a-16cf-4ea7-a27f-e1899ca9f9e8', 'serial': '8a71432a-16cf-4ea7-a27f-e1899ca9f9e8'} {{(pid=63088) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1046.284843] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e47f850-a25c-4d2f-bd34-69c0331348b8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.307033] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a67a1e8f-0e79-456e-8cc3-7bd0761e9bf5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.313839] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b96c16e-9ba1-43cf-8c3f-69ebe633ede1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.334701] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-489ef361-4999-46f7-989f-ec5e0099f5f2 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.350165] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-cf33486b-20b0-41d5-b03c-7534ec45d4ad tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] The volume has not been displaced from its original location: [datastore1] volume-8a71432a-16cf-4ea7-a27f-e1899ca9f9e8/volume-8a71432a-16cf-4ea7-a27f-e1899ca9f9e8.vmdk. No consolidation needed. {{(pid=63088) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1046.355634] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-cf33486b-20b0-41d5-b03c-7534ec45d4ad tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Reconfiguring VM instance instance-0000004d to detach disk 2001 {{(pid=63088) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1046.355998] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-62b94d0c-f017-46b5-a1a0-5d4fbacd07f1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.374059] env[63088]: DEBUG oslo_vmware.api [None req-cf33486b-20b0-41d5-b03c-7534ec45d4ad tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Waiting for the task: (returnval){ [ 1046.374059] env[63088]: value = "task-1285484" [ 1046.374059] env[63088]: _type = "Task" [ 1046.374059] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.383566] env[63088]: DEBUG oslo_vmware.api [None req-cf33486b-20b0-41d5-b03c-7534ec45d4ad tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285484, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.504228] env[63088]: DEBUG nova.virt.hardware [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1046.504936] env[63088]: DEBUG nova.virt.hardware [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1046.505252] env[63088]: DEBUG nova.virt.hardware [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1046.505607] env[63088]: DEBUG nova.virt.hardware [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1046.507032] env[63088]: DEBUG nova.virt.hardware [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1046.507032] env[63088]: DEBUG nova.virt.hardware [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1046.507032] env[63088]: DEBUG nova.virt.hardware [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1046.507032] env[63088]: DEBUG nova.virt.hardware [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1046.507032] env[63088]: DEBUG nova.virt.hardware [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1046.507032] env[63088]: DEBUG nova.virt.hardware [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1046.507032] env[63088]: DEBUG nova.virt.hardware [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1046.508028] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0250179-4b39-4c62-abf0-01ba3b16a9eb {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.516485] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29e0e960-2118-490f-9f93-6564bc25f23d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.531246] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] [instance: 9d2e0a07-3542-4e05-9902-82a949baa5a8] Instance VIF info [] {{(pid=63088) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1046.536989] env[63088]: DEBUG oslo.service.loopingcall [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1046.537316] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9d2e0a07-3542-4e05-9902-82a949baa5a8] Creating VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1046.537537] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9104d2e2-9f01-4b1b-ae48-07b8309cb978 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.554809] env[63088]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1046.554809] env[63088]: value = "task-1285485" [ 1046.554809] env[63088]: _type = "Task" [ 1046.554809] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.562306] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285485, 'name': CreateVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.589027] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Releasing lock "refresh_cache-71e37d8e-a454-46c4-a3cc-3d5671a32beb" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1046.589027] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Updated the network info_cache for instance {{(pid=63088) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 1046.589027] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1046.589027] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1046.589369] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1046.589369] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1046.589525] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1046.589803] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1046.589803] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63088) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1046.589919] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager.update_available_resource {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1046.884688] env[63088]: DEBUG oslo_vmware.api [None req-cf33486b-20b0-41d5-b03c-7534ec45d4ad tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285484, 'name': ReconfigVM_Task, 'duration_secs': 0.269465} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.884688] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-cf33486b-20b0-41d5-b03c-7534ec45d4ad tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Reconfigured VM instance instance-0000004d to detach disk 2001 {{(pid=63088) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1046.889192] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cb6cd7a3-324d-46d8-9db7-a14204380fd9 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.904882] env[63088]: DEBUG oslo_vmware.api [None req-cf33486b-20b0-41d5-b03c-7534ec45d4ad tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Waiting for the task: (returnval){ [ 1046.904882] env[63088]: value = "task-1285486" [ 1046.904882] env[63088]: _type = "Task" [ 1046.904882] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.912953] env[63088]: DEBUG oslo_vmware.api [None req-cf33486b-20b0-41d5-b03c-7534ec45d4ad tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285486, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.064153] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285485, 'name': CreateVM_Task, 'duration_secs': 0.262129} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.064337] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9d2e0a07-3542-4e05-9902-82a949baa5a8] Created VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1047.064736] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1047.064943] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1047.065293] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1047.065596] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-397e804c-e9cc-422d-b8aa-4a12e56f8c89 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.069765] env[63088]: DEBUG oslo_vmware.api [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Waiting for the task: (returnval){ [ 1047.069765] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]524156cb-8ed0-8a29-51cf-2985116c1130" [ 1047.069765] env[63088]: _type = "Task" [ 1047.069765] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.077202] env[63088]: DEBUG oslo_vmware.api [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]524156cb-8ed0-8a29-51cf-2985116c1130, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.093167] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1047.093360] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1047.093519] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1047.093671] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63088) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1047.094396] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33100995-77ba-416d-a12b-45ed1eb4d22d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.100941] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcaa115a-9c9c-4c7f-939c-beb8f730bcc5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.114217] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b3fdce9-305c-448c-b582-6cc93abd72dd {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.119352] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6637df9c-2998-4ad0-a1e4-2159a2020222 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.124245] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cb0cd7b-f64d-4c15-9885-c6d11a2906bc {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.140839] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Updating instance 'd94e57a3-743b-4bfe-b208-80f5748f8654' progress to 0 {{(pid=63088) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1047.169408] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180730MB free_disk=140GB free_vcpus=48 pci_devices=None {{(pid=63088) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1047.169584] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1047.169803] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1047.287201] env[63088]: DEBUG oslo_concurrency.lockutils [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquiring lock "b25b5883-20c0-4cc9-8cc1-2c22a0f5be81" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1047.287432] env[63088]: DEBUG oslo_concurrency.lockutils [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lock "b25b5883-20c0-4cc9-8cc1-2c22a0f5be81" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1047.414329] env[63088]: DEBUG oslo_vmware.api [None req-cf33486b-20b0-41d5-b03c-7534ec45d4ad tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285486, 'name': ReconfigVM_Task, 'duration_secs': 0.179725} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.414637] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-cf33486b-20b0-41d5-b03c-7534ec45d4ad tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-275932', 'volume_id': '8a71432a-16cf-4ea7-a27f-e1899ca9f9e8', 'name': 'volume-8a71432a-16cf-4ea7-a27f-e1899ca9f9e8', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '71e37d8e-a454-46c4-a3cc-3d5671a32beb', 'attached_at': '', 'detached_at': '', 'volume_id': '8a71432a-16cf-4ea7-a27f-e1899ca9f9e8', 'serial': '8a71432a-16cf-4ea7-a27f-e1899ca9f9e8'} {{(pid=63088) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1047.581478] env[63088]: DEBUG oslo_vmware.api [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]524156cb-8ed0-8a29-51cf-2985116c1130, 'name': SearchDatastore_Task, 'duration_secs': 0.008966} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.581478] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1047.581478] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] [instance: 9d2e0a07-3542-4e05-9902-82a949baa5a8] Processing image 1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1047.581478] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1047.581884] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1047.581884] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1047.582114] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b31d0e56-c123-4280-bd83-84f6a953e428 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.589975] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1047.590217] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1047.590950] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-83a85cc8-5794-43ed-a2b6-5df1b343422a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.595715] env[63088]: DEBUG oslo_vmware.api [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Waiting for the task: (returnval){ [ 1047.595715] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52bf4998-bb23-5892-9663-ef165712cbcb" [ 1047.595715] env[63088]: _type = "Task" [ 1047.595715] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.602851] env[63088]: DEBUG oslo_vmware.api [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52bf4998-bb23-5892-9663-ef165712cbcb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.646946] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1047.647302] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4731e29a-49f4-4346-bd98-96230696bd35 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.654047] env[63088]: DEBUG oslo_vmware.api [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 1047.654047] env[63088]: value = "task-1285487" [ 1047.654047] env[63088]: _type = "Task" [ 1047.654047] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.661862] env[63088]: DEBUG oslo_vmware.api [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285487, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.790231] env[63088]: DEBUG nova.compute.manager [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1047.960883] env[63088]: DEBUG nova.objects.instance [None req-cf33486b-20b0-41d5-b03c-7534ec45d4ad tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Lazy-loading 'flavor' on Instance uuid 71e37d8e-a454-46c4-a3cc-3d5671a32beb {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1048.109593] env[63088]: DEBUG oslo_vmware.api [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52bf4998-bb23-5892-9663-ef165712cbcb, 'name': SearchDatastore_Task, 'duration_secs': 0.008183} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.110476] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-41c41377-b34b-4666-80d8-69590700352d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.115625] env[63088]: DEBUG oslo_vmware.api [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Waiting for the task: (returnval){ [ 1048.115625] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]5211564a-e380-e468-c2bc-9609634cd84c" [ 1048.115625] env[63088]: _type = "Task" [ 1048.115625] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.122839] env[63088]: DEBUG oslo_vmware.api [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]5211564a-e380-e468-c2bc-9609634cd84c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.163126] env[63088]: DEBUG oslo_vmware.api [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285487, 'name': PowerOffVM_Task, 'duration_secs': 0.185244} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.163391] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1048.163587] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Updating instance 'd94e57a3-743b-4bfe-b208-80f5748f8654' progress to 17 {{(pid=63088) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1048.177045] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Applying migration context for instance d94e57a3-743b-4bfe-b208-80f5748f8654 as it has an incoming, in-progress migration 1873436d-a9d6-4d6c-80b5-d6b5decb0c77. Migration status is migrating {{(pid=63088) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1048.177430] env[63088]: INFO nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Updating resource usage from migration 1873436d-a9d6-4d6c-80b5-d6b5decb0c77 [ 1048.196348] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance 71e37d8e-a454-46c4-a3cc-3d5671a32beb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1048.196504] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance 2b87a9f1-5f10-43c2-8bc9-6d560ec88015 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1048.196641] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance 9d2e0a07-3542-4e05-9902-82a949baa5a8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1048.196762] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Migration 1873436d-a9d6-4d6c-80b5-d6b5decb0c77 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1048.196880] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance d94e57a3-743b-4bfe-b208-80f5748f8654 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1048.309773] env[63088]: DEBUG oslo_concurrency.lockutils [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1048.626119] env[63088]: DEBUG oslo_vmware.api [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]5211564a-e380-e468-c2bc-9609634cd84c, 'name': SearchDatastore_Task, 'duration_secs': 0.010391} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.626522] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1048.626640] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] 9d2e0a07-3542-4e05-9902-82a949baa5a8/9d2e0a07-3542-4e05-9902-82a949baa5a8.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1048.626894] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b7bb16a7-bb1f-43cb-b2f9-44299c852b1b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.633053] env[63088]: DEBUG oslo_vmware.api [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Waiting for the task: (returnval){ [ 1048.633053] env[63088]: value = "task-1285488" [ 1048.633053] env[63088]: _type = "Task" [ 1048.633053] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.641248] env[63088]: DEBUG oslo_vmware.api [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Task: {'id': task-1285488, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.669678] env[63088]: DEBUG nova.virt.hardware [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:33Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1048.669932] env[63088]: DEBUG nova.virt.hardware [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1048.670111] env[63088]: DEBUG nova.virt.hardware [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1048.670307] env[63088]: DEBUG nova.virt.hardware [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1048.670523] env[63088]: DEBUG nova.virt.hardware [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1048.670640] env[63088]: DEBUG nova.virt.hardware [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1048.670852] env[63088]: DEBUG nova.virt.hardware [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1048.671026] env[63088]: DEBUG nova.virt.hardware [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1048.671203] env[63088]: DEBUG nova.virt.hardware [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1048.671373] env[63088]: DEBUG nova.virt.hardware [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1048.671763] env[63088]: DEBUG nova.virt.hardware [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1048.676680] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c2db19df-5db3-41b0-8238-e9dde8ce5b34 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.691499] env[63088]: DEBUG oslo_vmware.api [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 1048.691499] env[63088]: value = "task-1285489" [ 1048.691499] env[63088]: _type = "Task" [ 1048.691499] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.699992] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance b25b5883-20c0-4cc9-8cc1-2c22a0f5be81 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1048.700242] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Total usable vcpus: 48, total allocated vcpus: 5 {{(pid=63088) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1048.700362] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1536MB phys_disk=200GB used_disk=5GB total_vcpus=48 used_vcpus=5 pci_stats=[] {{(pid=63088) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1048.702654] env[63088]: DEBUG oslo_vmware.api [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285489, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.785964] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa5310e9-c916-4ed5-b4a0-ecc1c43e2bed {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.794030] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68e2dff8-adfb-44bd-af83-12df83ad0714 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.827259] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7042dfa-dbd9-4b45-9d43-520a002dd177 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.835293] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5584b204-f93c-4199-97ff-1b6eec8afa91 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.849717] env[63088]: DEBUG nova.compute.provider_tree [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1048.969571] env[63088]: DEBUG oslo_concurrency.lockutils [None req-cf33486b-20b0-41d5-b03c-7534ec45d4ad tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Lock "71e37d8e-a454-46c4-a3cc-3d5671a32beb" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.223s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1049.143122] env[63088]: DEBUG oslo_vmware.api [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Task: {'id': task-1285488, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.451848} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.143398] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] 9d2e0a07-3542-4e05-9902-82a949baa5a8/9d2e0a07-3542-4e05-9902-82a949baa5a8.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1049.143620] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] [instance: 9d2e0a07-3542-4e05-9902-82a949baa5a8] Extending root virtual disk to 1048576 {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1049.143911] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c01316ee-dcda-4203-b351-7834a169e3ef {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.149882] env[63088]: DEBUG oslo_vmware.api [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Waiting for the task: (returnval){ [ 1049.149882] env[63088]: value = "task-1285490" [ 1049.149882] env[63088]: _type = "Task" [ 1049.149882] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.156933] env[63088]: DEBUG oslo_vmware.api [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Task: {'id': task-1285490, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.200800] env[63088]: DEBUG oslo_vmware.api [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285489, 'name': ReconfigVM_Task, 'duration_secs': 0.135714} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.201139] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Updating instance 'd94e57a3-743b-4bfe-b208-80f5748f8654' progress to 33 {{(pid=63088) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1049.352853] env[63088]: DEBUG nova.scheduler.client.report [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1049.662096] env[63088]: DEBUG oslo_vmware.api [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Task: {'id': task-1285490, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062473} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.662426] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] [instance: 9d2e0a07-3542-4e05-9902-82a949baa5a8] Extended root virtual disk {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1049.663296] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88f8daef-531f-40ac-9e4b-7603ea6547bc {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.691747] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] [instance: 9d2e0a07-3542-4e05-9902-82a949baa5a8] Reconfiguring VM instance instance-00000066 to attach disk [datastore1] 9d2e0a07-3542-4e05-9902-82a949baa5a8/9d2e0a07-3542-4e05-9902-82a949baa5a8.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1049.692323] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8b1feb70-a3ad-48f9-b81c-4037ce090ede {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.715536] env[63088]: DEBUG nova.virt.hardware [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1049.715761] env[63088]: DEBUG nova.virt.hardware [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1049.715922] env[63088]: DEBUG nova.virt.hardware [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1049.716124] env[63088]: DEBUG nova.virt.hardware [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1049.716278] env[63088]: DEBUG nova.virt.hardware [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1049.716430] env[63088]: DEBUG nova.virt.hardware [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1049.716635] env[63088]: DEBUG nova.virt.hardware [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1049.716802] env[63088]: DEBUG nova.virt.hardware [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1049.716971] env[63088]: DEBUG nova.virt.hardware [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1049.717154] env[63088]: DEBUG nova.virt.hardware [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1049.717334] env[63088]: DEBUG nova.virt.hardware [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1049.722322] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Reconfiguring VM instance instance-00000065 to detach disk 2000 {{(pid=63088) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1049.722603] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-88f0b97d-d654-4f33-ac31-7226c2bba752 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.741673] env[63088]: DEBUG oslo_vmware.api [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Waiting for the task: (returnval){ [ 1049.741673] env[63088]: value = "task-1285491" [ 1049.741673] env[63088]: _type = "Task" [ 1049.741673] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.742902] env[63088]: DEBUG oslo_vmware.api [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 1049.742902] env[63088]: value = "task-1285492" [ 1049.742902] env[63088]: _type = "Task" [ 1049.742902] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.754125] env[63088]: DEBUG oslo_vmware.api [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Task: {'id': task-1285491, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.756959] env[63088]: DEBUG oslo_vmware.api [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285492, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.857322] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63088) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1049.857628] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.688s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1049.858037] env[63088]: DEBUG oslo_concurrency.lockutils [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.548s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1049.859670] env[63088]: INFO nova.compute.claims [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1049.983614] env[63088]: DEBUG oslo_concurrency.lockutils [None req-06a926c1-b0ea-424d-a411-1ca268858b9f tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Acquiring lock "71e37d8e-a454-46c4-a3cc-3d5671a32beb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1049.983902] env[63088]: DEBUG oslo_concurrency.lockutils [None req-06a926c1-b0ea-424d-a411-1ca268858b9f tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Lock "71e37d8e-a454-46c4-a3cc-3d5671a32beb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1049.984164] env[63088]: DEBUG oslo_concurrency.lockutils [None req-06a926c1-b0ea-424d-a411-1ca268858b9f tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Acquiring lock "71e37d8e-a454-46c4-a3cc-3d5671a32beb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1049.984370] env[63088]: DEBUG oslo_concurrency.lockutils [None req-06a926c1-b0ea-424d-a411-1ca268858b9f tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Lock "71e37d8e-a454-46c4-a3cc-3d5671a32beb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1049.984563] env[63088]: DEBUG oslo_concurrency.lockutils [None req-06a926c1-b0ea-424d-a411-1ca268858b9f tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Lock "71e37d8e-a454-46c4-a3cc-3d5671a32beb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1049.987174] env[63088]: INFO nova.compute.manager [None req-06a926c1-b0ea-424d-a411-1ca268858b9f tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Terminating instance [ 1049.988821] env[63088]: DEBUG nova.compute.manager [None req-06a926c1-b0ea-424d-a411-1ca268858b9f tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1049.989046] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-06a926c1-b0ea-424d-a411-1ca268858b9f tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1049.989931] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0440da05-ea5d-4d84-987e-7b85debb82fd {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.997221] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-06a926c1-b0ea-424d-a411-1ca268858b9f tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1049.997455] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-82704753-43a8-4c80-9b34-9121eb43af1f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.003071] env[63088]: DEBUG oslo_vmware.api [None req-06a926c1-b0ea-424d-a411-1ca268858b9f tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Waiting for the task: (returnval){ [ 1050.003071] env[63088]: value = "task-1285493" [ 1050.003071] env[63088]: _type = "Task" [ 1050.003071] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.010452] env[63088]: DEBUG oslo_vmware.api [None req-06a926c1-b0ea-424d-a411-1ca268858b9f tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285493, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.255222] env[63088]: DEBUG oslo_vmware.api [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Task: {'id': task-1285491, 'name': ReconfigVM_Task, 'duration_secs': 0.329252} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.258047] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] [instance: 9d2e0a07-3542-4e05-9902-82a949baa5a8] Reconfigured VM instance instance-00000066 to attach disk [datastore1] 9d2e0a07-3542-4e05-9902-82a949baa5a8/9d2e0a07-3542-4e05-9902-82a949baa5a8.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1050.258713] env[63088]: DEBUG oslo_vmware.api [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285492, 'name': ReconfigVM_Task, 'duration_secs': 0.166053} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.258929] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9683a64f-30f1-4427-bf5c-3d81d1c638e4 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.260390] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Reconfigured VM instance instance-00000065 to detach disk 2000 {{(pid=63088) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1050.261121] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc8f6059-3f0b-4527-9e38-5e1f495080d5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.282615] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Reconfiguring VM instance instance-00000065 to attach disk [datastore1] d94e57a3-743b-4bfe-b208-80f5748f8654/d94e57a3-743b-4bfe-b208-80f5748f8654.vmdk or device None with type thin {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1050.283677] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-998da47c-5c37-40d9-8cfa-d442f7a14c25 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.299064] env[63088]: DEBUG oslo_vmware.api [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Waiting for the task: (returnval){ [ 1050.299064] env[63088]: value = "task-1285494" [ 1050.299064] env[63088]: _type = "Task" [ 1050.299064] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.304023] env[63088]: DEBUG oslo_vmware.api [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 1050.304023] env[63088]: value = "task-1285495" [ 1050.304023] env[63088]: _type = "Task" [ 1050.304023] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.307310] env[63088]: DEBUG oslo_vmware.api [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Task: {'id': task-1285494, 'name': Rename_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.314430] env[63088]: DEBUG oslo_vmware.api [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285495, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.513409] env[63088]: DEBUG oslo_vmware.api [None req-06a926c1-b0ea-424d-a411-1ca268858b9f tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285493, 'name': PowerOffVM_Task, 'duration_secs': 0.196057} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.513686] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-06a926c1-b0ea-424d-a411-1ca268858b9f tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1050.513859] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-06a926c1-b0ea-424d-a411-1ca268858b9f tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Unregistering the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1050.514156] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-aefd311d-44e9-4e99-86b1-35cdfc8574cd {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.576189] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-06a926c1-b0ea-424d-a411-1ca268858b9f tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Unregistered the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1050.576453] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-06a926c1-b0ea-424d-a411-1ca268858b9f tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Deleting contents of the VM from datastore datastore2 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1050.576639] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-06a926c1-b0ea-424d-a411-1ca268858b9f tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Deleting the datastore file [datastore2] 71e37d8e-a454-46c4-a3cc-3d5671a32beb {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1050.576923] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4d3cf2da-992d-4811-9be9-6c7fa79ee52d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.583358] env[63088]: DEBUG oslo_vmware.api [None req-06a926c1-b0ea-424d-a411-1ca268858b9f tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Waiting for the task: (returnval){ [ 1050.583358] env[63088]: value = "task-1285497" [ 1050.583358] env[63088]: _type = "Task" [ 1050.583358] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.590675] env[63088]: DEBUG oslo_vmware.api [None req-06a926c1-b0ea-424d-a411-1ca268858b9f tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285497, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.807728] env[63088]: DEBUG oslo_vmware.api [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Task: {'id': task-1285494, 'name': Rename_Task, 'duration_secs': 0.146135} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.810961] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] [instance: 9d2e0a07-3542-4e05-9902-82a949baa5a8] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1050.811283] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a99aea2f-0ada-4e63-847b-dd71d8e0d4d7 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.817300] env[63088]: DEBUG oslo_vmware.api [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285495, 'name': ReconfigVM_Task, 'duration_secs': 0.261075} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.818440] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Reconfigured VM instance instance-00000065 to attach disk [datastore1] d94e57a3-743b-4bfe-b208-80f5748f8654/d94e57a3-743b-4bfe-b208-80f5748f8654.vmdk or device None with type thin {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1050.818712] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Updating instance 'd94e57a3-743b-4bfe-b208-80f5748f8654' progress to 50 {{(pid=63088) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1050.821916] env[63088]: DEBUG oslo_vmware.api [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Waiting for the task: (returnval){ [ 1050.821916] env[63088]: value = "task-1285498" [ 1050.821916] env[63088]: _type = "Task" [ 1050.821916] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.829007] env[63088]: DEBUG oslo_vmware.api [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Task: {'id': task-1285498, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.952638] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1114e3c-3cf6-41d8-a69b-46d37f42ad57 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.962163] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-523489ee-8754-4c9f-bda6-2e17999cde70 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.992236] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d92e87d5-4f75-4baf-bae2-b19392639c36 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.999635] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7fe9575-990a-4653-a0b1-1a5888f3a6e9 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.013185] env[63088]: DEBUG nova.compute.provider_tree [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1051.093190] env[63088]: DEBUG oslo_vmware.api [None req-06a926c1-b0ea-424d-a411-1ca268858b9f tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285497, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.134521} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.093449] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-06a926c1-b0ea-424d-a411-1ca268858b9f tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1051.093672] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-06a926c1-b0ea-424d-a411-1ca268858b9f tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Deleted contents of the VM from datastore datastore2 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1051.093883] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-06a926c1-b0ea-424d-a411-1ca268858b9f tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1051.094078] env[63088]: INFO nova.compute.manager [None req-06a926c1-b0ea-424d-a411-1ca268858b9f tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1051.094329] env[63088]: DEBUG oslo.service.loopingcall [None req-06a926c1-b0ea-424d-a411-1ca268858b9f tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1051.094525] env[63088]: DEBUG nova.compute.manager [-] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1051.094621] env[63088]: DEBUG nova.network.neutron [-] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1051.325399] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-891d664b-d792-4950-949e-af16e0bdfa82 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.337117] env[63088]: DEBUG oslo_vmware.api [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Task: {'id': task-1285498, 'name': PowerOnVM_Task, 'duration_secs': 0.443639} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.350016] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] [instance: 9d2e0a07-3542-4e05-9902-82a949baa5a8] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1051.350236] env[63088]: DEBUG nova.compute.manager [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] [instance: 9d2e0a07-3542-4e05-9902-82a949baa5a8] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1051.351127] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01327a6d-9edf-4972-9e1e-f1018cd60a3f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.354087] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fadae087-5ac0-491f-aec5-bee676ade78e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.376813] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Updating instance 'd94e57a3-743b-4bfe-b208-80f5748f8654' progress to 67 {{(pid=63088) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1051.516247] env[63088]: DEBUG nova.scheduler.client.report [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1051.530364] env[63088]: DEBUG nova.compute.manager [req-f9aab420-bbb1-407d-9772-336f01a94e90 req-dae7508a-5b16-40f7-8fcc-f3b85dae5f6e service nova] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Received event network-vif-deleted-79ca8279-a3e2-4792-a4b5-57f899be5c41 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1051.530364] env[63088]: INFO nova.compute.manager [req-f9aab420-bbb1-407d-9772-336f01a94e90 req-dae7508a-5b16-40f7-8fcc-f3b85dae5f6e service nova] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Neutron deleted interface 79ca8279-a3e2-4792-a4b5-57f899be5c41; detaching it from the instance and deleting it from the info cache [ 1051.530364] env[63088]: DEBUG nova.network.neutron [req-f9aab420-bbb1-407d-9772-336f01a94e90 req-dae7508a-5b16-40f7-8fcc-f3b85dae5f6e service nova] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1051.889505] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1051.921947] env[63088]: DEBUG nova.network.neutron [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Port 024c5e7f-23a2-46df-83c1-5195ff662b68 binding to destination host cpu-1 is already ACTIVE {{(pid=63088) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1052.009702] env[63088]: DEBUG nova.network.neutron [-] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1052.020535] env[63088]: DEBUG oslo_concurrency.lockutils [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.162s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1052.021138] env[63088]: DEBUG nova.compute.manager [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1052.023977] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.134s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1052.023977] env[63088]: DEBUG nova.objects.instance [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] [instance: 9d2e0a07-3542-4e05-9902-82a949baa5a8] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63088) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1052.032969] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-de7bb610-7b43-4a7a-86b5-3fb0aa54bc4e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.044965] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a381ce65-8e9e-4285-81fb-fb25a66a764f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.073545] env[63088]: DEBUG nova.compute.manager [req-f9aab420-bbb1-407d-9772-336f01a94e90 req-dae7508a-5b16-40f7-8fcc-f3b85dae5f6e service nova] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Detach interface failed, port_id=79ca8279-a3e2-4792-a4b5-57f899be5c41, reason: Instance 71e37d8e-a454-46c4-a3cc-3d5671a32beb could not be found. {{(pid=63088) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1052.474987] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c63a21ad-0983-4a16-b773-f853089fd2f6 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Acquiring lock "9d2e0a07-3542-4e05-9902-82a949baa5a8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1052.475234] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c63a21ad-0983-4a16-b773-f853089fd2f6 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Lock "9d2e0a07-3542-4e05-9902-82a949baa5a8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1052.475436] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c63a21ad-0983-4a16-b773-f853089fd2f6 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Acquiring lock "9d2e0a07-3542-4e05-9902-82a949baa5a8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1052.475621] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c63a21ad-0983-4a16-b773-f853089fd2f6 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Lock "9d2e0a07-3542-4e05-9902-82a949baa5a8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1052.475794] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c63a21ad-0983-4a16-b773-f853089fd2f6 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Lock "9d2e0a07-3542-4e05-9902-82a949baa5a8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1052.477830] env[63088]: INFO nova.compute.manager [None req-c63a21ad-0983-4a16-b773-f853089fd2f6 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] [instance: 9d2e0a07-3542-4e05-9902-82a949baa5a8] Terminating instance [ 1052.479400] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c63a21ad-0983-4a16-b773-f853089fd2f6 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Acquiring lock "refresh_cache-9d2e0a07-3542-4e05-9902-82a949baa5a8" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1052.479607] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c63a21ad-0983-4a16-b773-f853089fd2f6 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Acquired lock "refresh_cache-9d2e0a07-3542-4e05-9902-82a949baa5a8" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1052.479801] env[63088]: DEBUG nova.network.neutron [None req-c63a21ad-0983-4a16-b773-f853089fd2f6 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] [instance: 9d2e0a07-3542-4e05-9902-82a949baa5a8] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1052.512666] env[63088]: INFO nova.compute.manager [-] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Took 1.42 seconds to deallocate network for instance. [ 1052.528466] env[63088]: DEBUG nova.compute.utils [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1052.532876] env[63088]: DEBUG nova.compute.manager [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1052.532876] env[63088]: DEBUG nova.network.neutron [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1052.570070] env[63088]: DEBUG nova.policy [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '787d7bd9f8e7490c9f24ff0e159bc98f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'dace8b5181b84623b08f903d12dfd31e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 1052.834394] env[63088]: DEBUG nova.network.neutron [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Successfully created port: f344e65c-4141-4c1d-a8e0-4aa1571a8e2f {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1052.943421] env[63088]: DEBUG oslo_concurrency.lockutils [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquiring lock "d94e57a3-743b-4bfe-b208-80f5748f8654-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1052.943955] env[63088]: DEBUG oslo_concurrency.lockutils [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "d94e57a3-743b-4bfe-b208-80f5748f8654-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1052.944127] env[63088]: DEBUG oslo_concurrency.lockutils [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "d94e57a3-743b-4bfe-b208-80f5748f8654-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1053.000948] env[63088]: DEBUG nova.network.neutron [None req-c63a21ad-0983-4a16-b773-f853089fd2f6 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] [instance: 9d2e0a07-3542-4e05-9902-82a949baa5a8] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1053.019348] env[63088]: DEBUG oslo_concurrency.lockutils [None req-06a926c1-b0ea-424d-a411-1ca268858b9f tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1053.032975] env[63088]: DEBUG nova.compute.manager [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1053.036643] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4de60e5d-3efd-4311-8e6e-e6e9074a6a59 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.013s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1053.038098] env[63088]: DEBUG oslo_concurrency.lockutils [None req-06a926c1-b0ea-424d-a411-1ca268858b9f tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.019s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1053.038325] env[63088]: DEBUG nova.objects.instance [None req-06a926c1-b0ea-424d-a411-1ca268858b9f tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Lazy-loading 'resources' on Instance uuid 71e37d8e-a454-46c4-a3cc-3d5671a32beb {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1053.080430] env[63088]: DEBUG nova.network.neutron [None req-c63a21ad-0983-4a16-b773-f853089fd2f6 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] [instance: 9d2e0a07-3542-4e05-9902-82a949baa5a8] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1053.583700] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c63a21ad-0983-4a16-b773-f853089fd2f6 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Releasing lock "refresh_cache-9d2e0a07-3542-4e05-9902-82a949baa5a8" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1053.584143] env[63088]: DEBUG nova.compute.manager [None req-c63a21ad-0983-4a16-b773-f853089fd2f6 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] [instance: 9d2e0a07-3542-4e05-9902-82a949baa5a8] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1053.584350] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-c63a21ad-0983-4a16-b773-f853089fd2f6 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] [instance: 9d2e0a07-3542-4e05-9902-82a949baa5a8] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1053.585249] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a739cfbd-aff7-41e2-bfb1-113195c8ae56 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.594144] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-c63a21ad-0983-4a16-b773-f853089fd2f6 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] [instance: 9d2e0a07-3542-4e05-9902-82a949baa5a8] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1053.594401] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3e983e33-c187-46cf-9bb5-1ff7f697475d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.603169] env[63088]: DEBUG oslo_vmware.api [None req-c63a21ad-0983-4a16-b773-f853089fd2f6 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Waiting for the task: (returnval){ [ 1053.603169] env[63088]: value = "task-1285499" [ 1053.603169] env[63088]: _type = "Task" [ 1053.603169] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.614238] env[63088]: DEBUG oslo_vmware.api [None req-c63a21ad-0983-4a16-b773-f853089fd2f6 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Task: {'id': task-1285499, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.654370] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e806d1a8-c83d-4f0b-8404-bf890d4bc08b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.661975] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55d4e516-bb70-4ca0-8a48-716e4a4aa828 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.693678] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6acef7ad-35f2-4990-9f50-a9c4efc09f0c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.701179] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c772244-a61c-4fdb-aae1-ea742f7849da {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.719628] env[63088]: DEBUG nova.compute.provider_tree [None req-06a926c1-b0ea-424d-a411-1ca268858b9f tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1053.987353] env[63088]: DEBUG oslo_concurrency.lockutils [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquiring lock "refresh_cache-d94e57a3-743b-4bfe-b208-80f5748f8654" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1053.987623] env[63088]: DEBUG oslo_concurrency.lockutils [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquired lock "refresh_cache-d94e57a3-743b-4bfe-b208-80f5748f8654" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1053.987868] env[63088]: DEBUG nova.network.neutron [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1054.047523] env[63088]: DEBUG nova.compute.manager [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1054.073012] env[63088]: DEBUG nova.virt.hardware [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1054.073297] env[63088]: DEBUG nova.virt.hardware [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1054.073460] env[63088]: DEBUG nova.virt.hardware [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1054.073648] env[63088]: DEBUG nova.virt.hardware [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1054.073802] env[63088]: DEBUG nova.virt.hardware [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1054.073996] env[63088]: DEBUG nova.virt.hardware [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1054.074320] env[63088]: DEBUG nova.virt.hardware [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1054.074494] env[63088]: DEBUG nova.virt.hardware [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1054.074667] env[63088]: DEBUG nova.virt.hardware [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1054.074835] env[63088]: DEBUG nova.virt.hardware [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1054.075024] env[63088]: DEBUG nova.virt.hardware [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1054.075874] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-043eb172-d37a-40fc-98b2-c67994ec0afb {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.083711] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c126474-0004-41f5-b07a-f916db3c6ec1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.113018] env[63088]: DEBUG oslo_vmware.api [None req-c63a21ad-0983-4a16-b773-f853089fd2f6 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Task: {'id': task-1285499, 'name': PowerOffVM_Task, 'duration_secs': 0.122744} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.113362] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-c63a21ad-0983-4a16-b773-f853089fd2f6 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] [instance: 9d2e0a07-3542-4e05-9902-82a949baa5a8] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1054.113575] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-c63a21ad-0983-4a16-b773-f853089fd2f6 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] [instance: 9d2e0a07-3542-4e05-9902-82a949baa5a8] Unregistering the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1054.113876] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5f399dbb-4e38-47c4-9ae3-9d2f5d217576 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.143197] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-c63a21ad-0983-4a16-b773-f853089fd2f6 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] [instance: 9d2e0a07-3542-4e05-9902-82a949baa5a8] Unregistered the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1054.143441] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-c63a21ad-0983-4a16-b773-f853089fd2f6 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] [instance: 9d2e0a07-3542-4e05-9902-82a949baa5a8] Deleting contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1054.143636] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-c63a21ad-0983-4a16-b773-f853089fd2f6 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Deleting the datastore file [datastore1] 9d2e0a07-3542-4e05-9902-82a949baa5a8 {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1054.143924] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fbb08c59-ae0d-4300-bb51-4af0a0f9875e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.150452] env[63088]: DEBUG oslo_vmware.api [None req-c63a21ad-0983-4a16-b773-f853089fd2f6 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Waiting for the task: (returnval){ [ 1054.150452] env[63088]: value = "task-1285501" [ 1054.150452] env[63088]: _type = "Task" [ 1054.150452] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.159540] env[63088]: DEBUG oslo_vmware.api [None req-c63a21ad-0983-4a16-b773-f853089fd2f6 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Task: {'id': task-1285501, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.223804] env[63088]: DEBUG nova.scheduler.client.report [None req-06a926c1-b0ea-424d-a411-1ca268858b9f tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1054.390233] env[63088]: DEBUG nova.network.neutron [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Successfully updated port: f344e65c-4141-4c1d-a8e0-4aa1571a8e2f {{(pid=63088) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1054.475152] env[63088]: DEBUG nova.compute.manager [req-aa356f61-b3ad-4ef9-9757-8eac18208d83 req-832102c2-9090-4f0b-91bc-f2c28c34f6ee service nova] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Received event network-vif-plugged-f344e65c-4141-4c1d-a8e0-4aa1571a8e2f {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1054.475507] env[63088]: DEBUG oslo_concurrency.lockutils [req-aa356f61-b3ad-4ef9-9757-8eac18208d83 req-832102c2-9090-4f0b-91bc-f2c28c34f6ee service nova] Acquiring lock "b25b5883-20c0-4cc9-8cc1-2c22a0f5be81-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1054.475675] env[63088]: DEBUG oslo_concurrency.lockutils [req-aa356f61-b3ad-4ef9-9757-8eac18208d83 req-832102c2-9090-4f0b-91bc-f2c28c34f6ee service nova] Lock "b25b5883-20c0-4cc9-8cc1-2c22a0f5be81-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1054.475766] env[63088]: DEBUG oslo_concurrency.lockutils [req-aa356f61-b3ad-4ef9-9757-8eac18208d83 req-832102c2-9090-4f0b-91bc-f2c28c34f6ee service nova] Lock "b25b5883-20c0-4cc9-8cc1-2c22a0f5be81-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1054.475975] env[63088]: DEBUG nova.compute.manager [req-aa356f61-b3ad-4ef9-9757-8eac18208d83 req-832102c2-9090-4f0b-91bc-f2c28c34f6ee service nova] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] No waiting events found dispatching network-vif-plugged-f344e65c-4141-4c1d-a8e0-4aa1571a8e2f {{(pid=63088) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1054.476128] env[63088]: WARNING nova.compute.manager [req-aa356f61-b3ad-4ef9-9757-8eac18208d83 req-832102c2-9090-4f0b-91bc-f2c28c34f6ee service nova] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Received unexpected event network-vif-plugged-f344e65c-4141-4c1d-a8e0-4aa1571a8e2f for instance with vm_state building and task_state spawning. [ 1054.659775] env[63088]: DEBUG oslo_vmware.api [None req-c63a21ad-0983-4a16-b773-f853089fd2f6 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Task: {'id': task-1285501, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.096342} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.660073] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-c63a21ad-0983-4a16-b773-f853089fd2f6 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1054.660291] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-c63a21ad-0983-4a16-b773-f853089fd2f6 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] [instance: 9d2e0a07-3542-4e05-9902-82a949baa5a8] Deleted contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1054.660482] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-c63a21ad-0983-4a16-b773-f853089fd2f6 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] [instance: 9d2e0a07-3542-4e05-9902-82a949baa5a8] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1054.660652] env[63088]: INFO nova.compute.manager [None req-c63a21ad-0983-4a16-b773-f853089fd2f6 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] [instance: 9d2e0a07-3542-4e05-9902-82a949baa5a8] Took 1.08 seconds to destroy the instance on the hypervisor. [ 1054.660896] env[63088]: DEBUG oslo.service.loopingcall [None req-c63a21ad-0983-4a16-b773-f853089fd2f6 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1054.661123] env[63088]: DEBUG nova.compute.manager [-] [instance: 9d2e0a07-3542-4e05-9902-82a949baa5a8] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1054.661223] env[63088]: DEBUG nova.network.neutron [-] [instance: 9d2e0a07-3542-4e05-9902-82a949baa5a8] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1054.680166] env[63088]: DEBUG nova.network.neutron [-] [instance: 9d2e0a07-3542-4e05-9902-82a949baa5a8] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1054.704740] env[63088]: DEBUG nova.network.neutron [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Updating instance_info_cache with network_info: [{"id": "024c5e7f-23a2-46df-83c1-5195ff662b68", "address": "fa:16:3e:76:26:9c", "network": {"id": "5b801381-6bb3-45cd-b81f-92b5d757e080", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1481457494-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e48f62dee9ad4e9b94b67bd871db5c63", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "051f343d-ac4f-4070-a26d-467603122c81", "external-id": "nsx-vlan-transportzone-277", "segmentation_id": 277, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap024c5e7f-23", "ovs_interfaceid": "024c5e7f-23a2-46df-83c1-5195ff662b68", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1054.728441] env[63088]: DEBUG oslo_concurrency.lockutils [None req-06a926c1-b0ea-424d-a411-1ca268858b9f tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.690s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1054.750562] env[63088]: INFO nova.scheduler.client.report [None req-06a926c1-b0ea-424d-a411-1ca268858b9f tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Deleted allocations for instance 71e37d8e-a454-46c4-a3cc-3d5671a32beb [ 1054.893070] env[63088]: DEBUG oslo_concurrency.lockutils [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquiring lock "refresh_cache-b25b5883-20c0-4cc9-8cc1-2c22a0f5be81" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1054.893372] env[63088]: DEBUG oslo_concurrency.lockutils [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquired lock "refresh_cache-b25b5883-20c0-4cc9-8cc1-2c22a0f5be81" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1054.893372] env[63088]: DEBUG nova.network.neutron [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1054.961479] env[63088]: DEBUG nova.compute.manager [None req-72711027-105f-439f-a56d-b2a885190dae tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1054.962598] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a7b70c8-51bd-4eab-ba60-53e39f46a567 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.183174] env[63088]: DEBUG nova.network.neutron [-] [instance: 9d2e0a07-3542-4e05-9902-82a949baa5a8] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1055.207340] env[63088]: DEBUG oslo_concurrency.lockutils [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Releasing lock "refresh_cache-d94e57a3-743b-4bfe-b208-80f5748f8654" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1055.258935] env[63088]: DEBUG oslo_concurrency.lockutils [None req-06a926c1-b0ea-424d-a411-1ca268858b9f tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Lock "71e37d8e-a454-46c4-a3cc-3d5671a32beb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.275s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1055.426739] env[63088]: DEBUG nova.network.neutron [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1055.473636] env[63088]: INFO nova.compute.manager [None req-72711027-105f-439f-a56d-b2a885190dae tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] instance snapshotting [ 1055.473978] env[63088]: DEBUG nova.objects.instance [None req-72711027-105f-439f-a56d-b2a885190dae tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lazy-loading 'flavor' on Instance uuid 2b87a9f1-5f10-43c2-8bc9-6d560ec88015 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1055.567667] env[63088]: DEBUG nova.network.neutron [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Updating instance_info_cache with network_info: [{"id": "f344e65c-4141-4c1d-a8e0-4aa1571a8e2f", "address": "fa:16:3e:3f:78:a4", "network": {"id": "dfcbe153-bc01-4362-9247-de9f0b1c847b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-841757397-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dace8b5181b84623b08f903d12dfd31e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c68b7663-4f0e-47f0-ac7f-40c6d952f7bb", "external-id": "nsx-vlan-transportzone-696", "segmentation_id": 696, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf344e65c-41", "ovs_interfaceid": "f344e65c-4141-4c1d-a8e0-4aa1571a8e2f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1055.686365] env[63088]: INFO nova.compute.manager [-] [instance: 9d2e0a07-3542-4e05-9902-82a949baa5a8] Took 1.02 seconds to deallocate network for instance. [ 1055.731530] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dcedda8-5a6d-4092-9a09-d3c86b94f1df {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.751906] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e096ea25-29e6-4874-9a7f-cde817bf3f5a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.759219] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Updating instance 'd94e57a3-743b-4bfe-b208-80f5748f8654' progress to 83 {{(pid=63088) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1055.982435] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c0a29de-a173-40bf-929c-400e64a8d59a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.001965] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-002d83ed-24bd-4c8e-bfd3-6640258d811d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.069966] env[63088]: DEBUG oslo_concurrency.lockutils [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Releasing lock "refresh_cache-b25b5883-20c0-4cc9-8cc1-2c22a0f5be81" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1056.070313] env[63088]: DEBUG nova.compute.manager [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Instance network_info: |[{"id": "f344e65c-4141-4c1d-a8e0-4aa1571a8e2f", "address": "fa:16:3e:3f:78:a4", "network": {"id": "dfcbe153-bc01-4362-9247-de9f0b1c847b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-841757397-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dace8b5181b84623b08f903d12dfd31e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c68b7663-4f0e-47f0-ac7f-40c6d952f7bb", "external-id": "nsx-vlan-transportzone-696", "segmentation_id": 696, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf344e65c-41", "ovs_interfaceid": "f344e65c-4141-4c1d-a8e0-4aa1571a8e2f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1056.070760] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3f:78:a4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c68b7663-4f0e-47f0-ac7f-40c6d952f7bb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f344e65c-4141-4c1d-a8e0-4aa1571a8e2f', 'vif_model': 'vmxnet3'}] {{(pid=63088) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1056.078456] env[63088]: DEBUG oslo.service.loopingcall [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1056.079326] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Creating VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1056.079528] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c1491287-10ad-46fe-9c63-b5490bdf7818 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.099342] env[63088]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1056.099342] env[63088]: value = "task-1285502" [ 1056.099342] env[63088]: _type = "Task" [ 1056.099342] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.107184] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285502, 'name': CreateVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.192603] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c63a21ad-0983-4a16-b773-f853089fd2f6 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1056.192928] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c63a21ad-0983-4a16-b773-f853089fd2f6 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1056.193184] env[63088]: DEBUG nova.objects.instance [None req-c63a21ad-0983-4a16-b773-f853089fd2f6 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Lazy-loading 'resources' on Instance uuid 9d2e0a07-3542-4e05-9902-82a949baa5a8 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1056.266519] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1056.267290] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c211638c-917f-4cef-a79c-9b7dbaba7247 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.274580] env[63088]: DEBUG oslo_vmware.api [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 1056.274580] env[63088]: value = "task-1285503" [ 1056.274580] env[63088]: _type = "Task" [ 1056.274580] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.282837] env[63088]: DEBUG oslo_vmware.api [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285503, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.512834] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-72711027-105f-439f-a56d-b2a885190dae tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Creating Snapshot of the VM instance {{(pid=63088) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1056.513235] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-aeb26305-b37b-45ab-a0f7-c6551ce93b8f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.521054] env[63088]: DEBUG oslo_vmware.api [None req-72711027-105f-439f-a56d-b2a885190dae tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1056.521054] env[63088]: value = "task-1285504" [ 1056.521054] env[63088]: _type = "Task" [ 1056.521054] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.528922] env[63088]: DEBUG nova.compute.manager [req-95d36ec7-1895-4806-8cef-2aae9b8c4fc8 req-76a552ee-6171-4878-a17f-e20473a43c32 service nova] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Received event network-changed-f344e65c-4141-4c1d-a8e0-4aa1571a8e2f {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1056.529334] env[63088]: DEBUG nova.compute.manager [req-95d36ec7-1895-4806-8cef-2aae9b8c4fc8 req-76a552ee-6171-4878-a17f-e20473a43c32 service nova] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Refreshing instance network info cache due to event network-changed-f344e65c-4141-4c1d-a8e0-4aa1571a8e2f. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1056.529449] env[63088]: DEBUG oslo_concurrency.lockutils [req-95d36ec7-1895-4806-8cef-2aae9b8c4fc8 req-76a552ee-6171-4878-a17f-e20473a43c32 service nova] Acquiring lock "refresh_cache-b25b5883-20c0-4cc9-8cc1-2c22a0f5be81" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1056.529613] env[63088]: DEBUG oslo_concurrency.lockutils [req-95d36ec7-1895-4806-8cef-2aae9b8c4fc8 req-76a552ee-6171-4878-a17f-e20473a43c32 service nova] Acquired lock "refresh_cache-b25b5883-20c0-4cc9-8cc1-2c22a0f5be81" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1056.529795] env[63088]: DEBUG nova.network.neutron [req-95d36ec7-1895-4806-8cef-2aae9b8c4fc8 req-76a552ee-6171-4878-a17f-e20473a43c32 service nova] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Refreshing network info cache for port f344e65c-4141-4c1d-a8e0-4aa1571a8e2f {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1056.534408] env[63088]: DEBUG oslo_vmware.api [None req-72711027-105f-439f-a56d-b2a885190dae tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285504, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.610304] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285502, 'name': CreateVM_Task, 'duration_secs': 0.33624} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.611357] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Created VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1056.611357] env[63088]: DEBUG oslo_concurrency.lockutils [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1056.612554] env[63088]: DEBUG oslo_concurrency.lockutils [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1056.612554] env[63088]: DEBUG oslo_concurrency.lockutils [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1056.612554] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0b26d11e-8c00-4835-b6d4-a0c50495a32e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.616998] env[63088]: DEBUG oslo_vmware.api [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for the task: (returnval){ [ 1056.616998] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52d80851-caf4-41e2-b7f7-d703a5575f99" [ 1056.616998] env[63088]: _type = "Task" [ 1056.616998] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.627599] env[63088]: DEBUG oslo_vmware.api [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52d80851-caf4-41e2-b7f7-d703a5575f99, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.773777] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27d08622-a56c-4daf-a38d-bba1a7de2f91 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.786084] env[63088]: DEBUG oslo_vmware.api [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285503, 'name': PowerOnVM_Task, 'duration_secs': 0.39707} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.788062] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1056.788327] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-43cd68e9-20a1-467c-8c52-81726236fbc4 tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Updating instance 'd94e57a3-743b-4bfe-b208-80f5748f8654' progress to 100 {{(pid=63088) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1056.792831] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bb4cf13-c9df-425b-9362-c0636674e2cf {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.823642] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd80fdd8-5d71-452a-aebc-e244ca0e410d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.830913] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e43075b-fbf6-4709-a86b-6e27a9f41e82 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.844638] env[63088]: DEBUG nova.compute.provider_tree [None req-c63a21ad-0983-4a16-b773-f853089fd2f6 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1057.030821] env[63088]: DEBUG oslo_vmware.api [None req-72711027-105f-439f-a56d-b2a885190dae tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285504, 'name': CreateSnapshot_Task, 'duration_secs': 0.47143} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.030981] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-72711027-105f-439f-a56d-b2a885190dae tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Created Snapshot of the VM instance {{(pid=63088) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1057.031746] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5b13551-c336-4f2d-b81e-c60d646c1f9f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.130612] env[63088]: DEBUG oslo_vmware.api [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52d80851-caf4-41e2-b7f7-d703a5575f99, 'name': SearchDatastore_Task, 'duration_secs': 0.009421} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.132177] env[63088]: DEBUG oslo_concurrency.lockutils [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1057.132415] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Processing image 1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1057.132666] env[63088]: DEBUG oslo_concurrency.lockutils [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1057.132824] env[63088]: DEBUG oslo_concurrency.lockutils [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1057.133015] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1057.135787] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-76dab5d3-6ae7-490e-a893-86ba13bcbe17 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.137788] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Acquiring lock "3fa11db4-2f70-495e-a562-ac5f06153847" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1057.137999] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Lock "3fa11db4-2f70-495e-a562-ac5f06153847" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1057.147684] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1057.147869] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1057.149056] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3668e103-8db2-4a8a-a92b-df5924cf697f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.156690] env[63088]: DEBUG oslo_vmware.api [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for the task: (returnval){ [ 1057.156690] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]526e8531-e7b0-f3fd-b046-30c316349374" [ 1057.156690] env[63088]: _type = "Task" [ 1057.156690] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.165489] env[63088]: DEBUG oslo_vmware.api [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]526e8531-e7b0-f3fd-b046-30c316349374, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.254086] env[63088]: DEBUG nova.network.neutron [req-95d36ec7-1895-4806-8cef-2aae9b8c4fc8 req-76a552ee-6171-4878-a17f-e20473a43c32 service nova] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Updated VIF entry in instance network info cache for port f344e65c-4141-4c1d-a8e0-4aa1571a8e2f. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1057.254465] env[63088]: DEBUG nova.network.neutron [req-95d36ec7-1895-4806-8cef-2aae9b8c4fc8 req-76a552ee-6171-4878-a17f-e20473a43c32 service nova] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Updating instance_info_cache with network_info: [{"id": "f344e65c-4141-4c1d-a8e0-4aa1571a8e2f", "address": "fa:16:3e:3f:78:a4", "network": {"id": "dfcbe153-bc01-4362-9247-de9f0b1c847b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-841757397-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dace8b5181b84623b08f903d12dfd31e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c68b7663-4f0e-47f0-ac7f-40c6d952f7bb", "external-id": "nsx-vlan-transportzone-696", "segmentation_id": 696, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf344e65c-41", "ovs_interfaceid": "f344e65c-4141-4c1d-a8e0-4aa1571a8e2f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1057.347294] env[63088]: DEBUG nova.scheduler.client.report [None req-c63a21ad-0983-4a16-b773-f853089fd2f6 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1057.550888] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-72711027-105f-439f-a56d-b2a885190dae tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Creating linked-clone VM from snapshot {{(pid=63088) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1057.551296] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-23527012-6568-4da5-88e1-5c5242a79218 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.560232] env[63088]: DEBUG oslo_vmware.api [None req-72711027-105f-439f-a56d-b2a885190dae tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1057.560232] env[63088]: value = "task-1285505" [ 1057.560232] env[63088]: _type = "Task" [ 1057.560232] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.568468] env[63088]: DEBUG oslo_vmware.api [None req-72711027-105f-439f-a56d-b2a885190dae tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285505, 'name': CloneVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.642802] env[63088]: DEBUG nova.compute.manager [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1057.667791] env[63088]: DEBUG oslo_vmware.api [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]526e8531-e7b0-f3fd-b046-30c316349374, 'name': SearchDatastore_Task, 'duration_secs': 0.007945} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.668766] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-42c78605-ba02-412c-812c-268a211396ba {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.675151] env[63088]: DEBUG oslo_vmware.api [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for the task: (returnval){ [ 1057.675151] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]528b4038-ae68-e928-a074-12b607348078" [ 1057.675151] env[63088]: _type = "Task" [ 1057.675151] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.684516] env[63088]: DEBUG oslo_vmware.api [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]528b4038-ae68-e928-a074-12b607348078, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.756974] env[63088]: DEBUG oslo_concurrency.lockutils [req-95d36ec7-1895-4806-8cef-2aae9b8c4fc8 req-76a552ee-6171-4878-a17f-e20473a43c32 service nova] Releasing lock "refresh_cache-b25b5883-20c0-4cc9-8cc1-2c22a0f5be81" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1057.851753] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c63a21ad-0983-4a16-b773-f853089fd2f6 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.659s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1057.873454] env[63088]: INFO nova.scheduler.client.report [None req-c63a21ad-0983-4a16-b773-f853089fd2f6 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Deleted allocations for instance 9d2e0a07-3542-4e05-9902-82a949baa5a8 [ 1058.070408] env[63088]: DEBUG oslo_vmware.api [None req-72711027-105f-439f-a56d-b2a885190dae tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285505, 'name': CloneVM_Task} progress is 94%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.161552] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1058.161956] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1058.163667] env[63088]: INFO nova.compute.claims [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1058.185645] env[63088]: DEBUG oslo_vmware.api [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]528b4038-ae68-e928-a074-12b607348078, 'name': SearchDatastore_Task, 'duration_secs': 0.01266} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.185877] env[63088]: DEBUG oslo_concurrency.lockutils [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1058.186149] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore2] b25b5883-20c0-4cc9-8cc1-2c22a0f5be81/b25b5883-20c0-4cc9-8cc1-2c22a0f5be81.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1058.186404] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7b766cc7-09bf-473e-8626-6baee6978799 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.193432] env[63088]: DEBUG oslo_vmware.api [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for the task: (returnval){ [ 1058.193432] env[63088]: value = "task-1285506" [ 1058.193432] env[63088]: _type = "Task" [ 1058.193432] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.200773] env[63088]: DEBUG oslo_vmware.api [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285506, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.381124] env[63088]: DEBUG oslo_concurrency.lockutils [None req-c63a21ad-0983-4a16-b773-f853089fd2f6 tempest-ServerShowV254Test-1105322002 tempest-ServerShowV254Test-1105322002-project-member] Lock "9d2e0a07-3542-4e05-9902-82a949baa5a8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.906s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1058.572537] env[63088]: DEBUG oslo_vmware.api [None req-72711027-105f-439f-a56d-b2a885190dae tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285505, 'name': CloneVM_Task} progress is 100%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.704122] env[63088]: DEBUG oslo_vmware.api [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285506, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.45814} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.704489] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore2] b25b5883-20c0-4cc9-8cc1-2c22a0f5be81/b25b5883-20c0-4cc9-8cc1-2c22a0f5be81.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1058.704769] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Extending root virtual disk to 1048576 {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1058.705089] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d7ab311f-1978-43b7-9693-94d42b61957b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.712237] env[63088]: DEBUG oslo_vmware.api [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for the task: (returnval){ [ 1058.712237] env[63088]: value = "task-1285507" [ 1058.712237] env[63088]: _type = "Task" [ 1058.712237] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.720223] env[63088]: DEBUG oslo_vmware.api [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285507, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.071148] env[63088]: DEBUG oslo_vmware.api [None req-72711027-105f-439f-a56d-b2a885190dae tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285505, 'name': CloneVM_Task, 'duration_secs': 1.039473} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.071426] env[63088]: INFO nova.virt.vmwareapi.vmops [None req-72711027-105f-439f-a56d-b2a885190dae tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Created linked-clone VM from snapshot [ 1059.072248] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04679ca7-4a36-44fa-a691-20f39c802c3e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.079933] env[63088]: DEBUG nova.virt.vmwareapi.images [None req-72711027-105f-439f-a56d-b2a885190dae tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Uploading image 7862f3ca-7713-4ff8-8498-b50e9658c73f {{(pid=63088) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1059.103811] env[63088]: DEBUG oslo_vmware.rw_handles [None req-72711027-105f-439f-a56d-b2a885190dae tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1059.103811] env[63088]: value = "vm-275958" [ 1059.103811] env[63088]: _type = "VirtualMachine" [ 1059.103811] env[63088]: }. {{(pid=63088) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1059.104141] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-9e293698-495b-471a-a397-cf6bcb1e04b0 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.111855] env[63088]: DEBUG oslo_vmware.rw_handles [None req-72711027-105f-439f-a56d-b2a885190dae tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lease: (returnval){ [ 1059.111855] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]525dd64f-e1e6-a68f-df79-0426b123db7b" [ 1059.111855] env[63088]: _type = "HttpNfcLease" [ 1059.111855] env[63088]: } obtained for exporting VM: (result){ [ 1059.111855] env[63088]: value = "vm-275958" [ 1059.111855] env[63088]: _type = "VirtualMachine" [ 1059.111855] env[63088]: }. {{(pid=63088) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1059.112178] env[63088]: DEBUG oslo_vmware.api [None req-72711027-105f-439f-a56d-b2a885190dae tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the lease: (returnval){ [ 1059.112178] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]525dd64f-e1e6-a68f-df79-0426b123db7b" [ 1059.112178] env[63088]: _type = "HttpNfcLease" [ 1059.112178] env[63088]: } to be ready. {{(pid=63088) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1059.118783] env[63088]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1059.118783] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]525dd64f-e1e6-a68f-df79-0426b123db7b" [ 1059.118783] env[63088]: _type = "HttpNfcLease" [ 1059.118783] env[63088]: } is initializing. {{(pid=63088) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1059.224318] env[63088]: DEBUG oslo_vmware.api [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285507, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073074} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.224592] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Extended root virtual disk {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1059.225409] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c479e385-2853-4f57-adc4-95b9a3bcb295 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.248031] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Reconfiguring VM instance instance-00000067 to attach disk [datastore2] b25b5883-20c0-4cc9-8cc1-2c22a0f5be81/b25b5883-20c0-4cc9-8cc1-2c22a0f5be81.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1059.250721] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-440dd901-6aad-44b2-a884-8b6ddc89804c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.269957] env[63088]: DEBUG oslo_vmware.api [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for the task: (returnval){ [ 1059.269957] env[63088]: value = "task-1285509" [ 1059.269957] env[63088]: _type = "Task" [ 1059.269957] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.274089] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d768c8f4-470b-4d08-b47a-24629e8b6c31 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.283556] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22492f78-1fb7-4cbd-88f3-d7190a417a36 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.286609] env[63088]: DEBUG oslo_vmware.api [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285509, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.315526] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3b41880-c3a4-48f0-9e0b-17c2273481e0 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.323307] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0ea3df6-afdb-44a5-82d2-07c8a7264a05 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.336629] env[63088]: DEBUG nova.compute.provider_tree [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1059.357071] env[63088]: DEBUG oslo_concurrency.lockutils [None req-8044d1f3-cb0e-48b3-830d-fbfd6377d70c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquiring lock "d94e57a3-743b-4bfe-b208-80f5748f8654" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1059.357312] env[63088]: DEBUG oslo_concurrency.lockutils [None req-8044d1f3-cb0e-48b3-830d-fbfd6377d70c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "d94e57a3-743b-4bfe-b208-80f5748f8654" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1059.357499] env[63088]: DEBUG nova.compute.manager [None req-8044d1f3-cb0e-48b3-830d-fbfd6377d70c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Going to confirm migration 5 {{(pid=63088) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1059.621738] env[63088]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1059.621738] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]525dd64f-e1e6-a68f-df79-0426b123db7b" [ 1059.621738] env[63088]: _type = "HttpNfcLease" [ 1059.621738] env[63088]: } is ready. {{(pid=63088) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1059.622264] env[63088]: DEBUG oslo_vmware.rw_handles [None req-72711027-105f-439f-a56d-b2a885190dae tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1059.622264] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]525dd64f-e1e6-a68f-df79-0426b123db7b" [ 1059.622264] env[63088]: _type = "HttpNfcLease" [ 1059.622264] env[63088]: }. {{(pid=63088) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1059.622784] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90a7cfb2-02ae-4471-9219-e856b39ead7e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.629435] env[63088]: DEBUG oslo_vmware.rw_handles [None req-72711027-105f-439f-a56d-b2a885190dae tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52083c3a-9df7-e61e-1da8-93e0f94c74d7/disk-0.vmdk from lease info. {{(pid=63088) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1059.629639] env[63088]: DEBUG oslo_vmware.rw_handles [None req-72711027-105f-439f-a56d-b2a885190dae tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52083c3a-9df7-e61e-1da8-93e0f94c74d7/disk-0.vmdk for reading. {{(pid=63088) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1059.719369] env[63088]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-44ca45b5-640c-496c-8480-90e34180b5f9 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.779386] env[63088]: DEBUG oslo_vmware.api [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285509, 'name': ReconfigVM_Task, 'duration_secs': 0.317925} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.779729] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Reconfigured VM instance instance-00000067 to attach disk [datastore2] b25b5883-20c0-4cc9-8cc1-2c22a0f5be81/b25b5883-20c0-4cc9-8cc1-2c22a0f5be81.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1059.780372] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c70c6f79-b533-4628-be42-86bb2ce9ae29 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.786293] env[63088]: DEBUG oslo_vmware.api [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for the task: (returnval){ [ 1059.786293] env[63088]: value = "task-1285510" [ 1059.786293] env[63088]: _type = "Task" [ 1059.786293] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.793593] env[63088]: DEBUG oslo_vmware.api [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285510, 'name': Rename_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.839836] env[63088]: DEBUG nova.scheduler.client.report [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1059.907788] env[63088]: DEBUG oslo_concurrency.lockutils [None req-8044d1f3-cb0e-48b3-830d-fbfd6377d70c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquiring lock "refresh_cache-d94e57a3-743b-4bfe-b208-80f5748f8654" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1059.908111] env[63088]: DEBUG oslo_concurrency.lockutils [None req-8044d1f3-cb0e-48b3-830d-fbfd6377d70c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquired lock "refresh_cache-d94e57a3-743b-4bfe-b208-80f5748f8654" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1059.908339] env[63088]: DEBUG nova.network.neutron [None req-8044d1f3-cb0e-48b3-830d-fbfd6377d70c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1059.908560] env[63088]: DEBUG nova.objects.instance [None req-8044d1f3-cb0e-48b3-830d-fbfd6377d70c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lazy-loading 'info_cache' on Instance uuid d94e57a3-743b-4bfe-b208-80f5748f8654 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1060.298222] env[63088]: DEBUG oslo_vmware.api [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285510, 'name': Rename_Task, 'duration_secs': 0.135684} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.298586] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1060.298924] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a6ac1f51-3fe5-4443-8954-5a27e8cdaf1b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.305942] env[63088]: DEBUG oslo_vmware.api [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for the task: (returnval){ [ 1060.305942] env[63088]: value = "task-1285511" [ 1060.305942] env[63088]: _type = "Task" [ 1060.305942] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.313933] env[63088]: DEBUG oslo_vmware.api [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285511, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.345125] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.183s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1060.345792] env[63088]: DEBUG nova.compute.manager [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1060.816560] env[63088]: DEBUG oslo_vmware.api [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285511, 'name': PowerOnVM_Task, 'duration_secs': 0.449915} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.816947] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1060.817165] env[63088]: INFO nova.compute.manager [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Took 6.77 seconds to spawn the instance on the hypervisor. [ 1060.817478] env[63088]: DEBUG nova.compute.manager [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1060.818480] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4db48eda-b5f0-44c5-8c68-976ac761f536 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.851013] env[63088]: DEBUG nova.compute.utils [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1060.852682] env[63088]: DEBUG nova.compute.manager [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1060.853724] env[63088]: DEBUG nova.network.neutron [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1060.909532] env[63088]: DEBUG nova.policy [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd76b45ac31364495b86fad66b9acfb8a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2b18ddddf2314d83addf550b8cb91977', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 1061.314160] env[63088]: DEBUG nova.network.neutron [None req-8044d1f3-cb0e-48b3-830d-fbfd6377d70c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Updating instance_info_cache with network_info: [{"id": "024c5e7f-23a2-46df-83c1-5195ff662b68", "address": "fa:16:3e:76:26:9c", "network": {"id": "5b801381-6bb3-45cd-b81f-92b5d757e080", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1481457494-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e48f62dee9ad4e9b94b67bd871db5c63", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "051f343d-ac4f-4070-a26d-467603122c81", "external-id": "nsx-vlan-transportzone-277", "segmentation_id": 277, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap024c5e7f-23", "ovs_interfaceid": "024c5e7f-23a2-46df-83c1-5195ff662b68", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1061.334564] env[63088]: INFO nova.compute.manager [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Took 13.04 seconds to build instance. [ 1061.355573] env[63088]: DEBUG nova.compute.manager [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1061.400074] env[63088]: DEBUG nova.network.neutron [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Successfully created port: eb4d6233-2d9d-4756-8b9d-084bdfaeb3f7 {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1061.816879] env[63088]: DEBUG oslo_concurrency.lockutils [None req-8044d1f3-cb0e-48b3-830d-fbfd6377d70c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Releasing lock "refresh_cache-d94e57a3-743b-4bfe-b208-80f5748f8654" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1061.817223] env[63088]: DEBUG nova.objects.instance [None req-8044d1f3-cb0e-48b3-830d-fbfd6377d70c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lazy-loading 'migration_context' on Instance uuid d94e57a3-743b-4bfe-b208-80f5748f8654 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1061.837214] env[63088]: DEBUG oslo_concurrency.lockutils [None req-12baed78-9377-4a8a-a1cd-c5f9f4acfb51 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lock "b25b5883-20c0-4cc9-8cc1-2c22a0f5be81" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.550s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1061.985648] env[63088]: DEBUG nova.compute.manager [req-f215eb93-fa6e-4d54-ba13-e230bb080897 req-7616d55c-1c7a-423d-a689-67db4465b0e1 service nova] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Received event network-changed-f344e65c-4141-4c1d-a8e0-4aa1571a8e2f {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1061.985648] env[63088]: DEBUG nova.compute.manager [req-f215eb93-fa6e-4d54-ba13-e230bb080897 req-7616d55c-1c7a-423d-a689-67db4465b0e1 service nova] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Refreshing instance network info cache due to event network-changed-f344e65c-4141-4c1d-a8e0-4aa1571a8e2f. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1061.985947] env[63088]: DEBUG oslo_concurrency.lockutils [req-f215eb93-fa6e-4d54-ba13-e230bb080897 req-7616d55c-1c7a-423d-a689-67db4465b0e1 service nova] Acquiring lock "refresh_cache-b25b5883-20c0-4cc9-8cc1-2c22a0f5be81" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1061.986769] env[63088]: DEBUG oslo_concurrency.lockutils [req-f215eb93-fa6e-4d54-ba13-e230bb080897 req-7616d55c-1c7a-423d-a689-67db4465b0e1 service nova] Acquired lock "refresh_cache-b25b5883-20c0-4cc9-8cc1-2c22a0f5be81" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1061.986949] env[63088]: DEBUG nova.network.neutron [req-f215eb93-fa6e-4d54-ba13-e230bb080897 req-7616d55c-1c7a-423d-a689-67db4465b0e1 service nova] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Refreshing network info cache for port f344e65c-4141-4c1d-a8e0-4aa1571a8e2f {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1062.320108] env[63088]: DEBUG nova.objects.base [None req-8044d1f3-cb0e-48b3-830d-fbfd6377d70c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=63088) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1062.321165] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42c1d1f9-7eff-47f9-a09a-479b2faad715 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.343076] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9634e25a-e518-4aa3-9bb7-9f49bc75ef6a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.348804] env[63088]: DEBUG oslo_vmware.api [None req-8044d1f3-cb0e-48b3-830d-fbfd6377d70c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 1062.348804] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]521967a9-424d-71d4-0eb0-f54ae0989d1d" [ 1062.348804] env[63088]: _type = "Task" [ 1062.348804] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.357211] env[63088]: DEBUG oslo_vmware.api [None req-8044d1f3-cb0e-48b3-830d-fbfd6377d70c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]521967a9-424d-71d4-0eb0-f54ae0989d1d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.368572] env[63088]: DEBUG nova.compute.manager [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1062.389524] env[63088]: DEBUG nova.virt.hardware [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1062.389763] env[63088]: DEBUG nova.virt.hardware [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1062.389930] env[63088]: DEBUG nova.virt.hardware [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1062.390154] env[63088]: DEBUG nova.virt.hardware [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1062.390326] env[63088]: DEBUG nova.virt.hardware [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1062.390483] env[63088]: DEBUG nova.virt.hardware [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1062.390719] env[63088]: DEBUG nova.virt.hardware [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1062.390890] env[63088]: DEBUG nova.virt.hardware [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1062.391071] env[63088]: DEBUG nova.virt.hardware [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1062.391251] env[63088]: DEBUG nova.virt.hardware [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1062.391428] env[63088]: DEBUG nova.virt.hardware [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1062.392341] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-335c8969-e73d-4872-9409-22ef6a799c46 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.400642] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e97c71b5-61fa-4d97-abde-b0c710363fa9 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.777207] env[63088]: DEBUG nova.network.neutron [req-f215eb93-fa6e-4d54-ba13-e230bb080897 req-7616d55c-1c7a-423d-a689-67db4465b0e1 service nova] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Updated VIF entry in instance network info cache for port f344e65c-4141-4c1d-a8e0-4aa1571a8e2f. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1062.777661] env[63088]: DEBUG nova.network.neutron [req-f215eb93-fa6e-4d54-ba13-e230bb080897 req-7616d55c-1c7a-423d-a689-67db4465b0e1 service nova] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Updating instance_info_cache with network_info: [{"id": "f344e65c-4141-4c1d-a8e0-4aa1571a8e2f", "address": "fa:16:3e:3f:78:a4", "network": {"id": "dfcbe153-bc01-4362-9247-de9f0b1c847b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-841757397-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.152", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dace8b5181b84623b08f903d12dfd31e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c68b7663-4f0e-47f0-ac7f-40c6d952f7bb", "external-id": "nsx-vlan-transportzone-696", "segmentation_id": 696, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf344e65c-41", "ovs_interfaceid": "f344e65c-4141-4c1d-a8e0-4aa1571a8e2f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1062.861299] env[63088]: DEBUG oslo_vmware.api [None req-8044d1f3-cb0e-48b3-830d-fbfd6377d70c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]521967a9-424d-71d4-0eb0-f54ae0989d1d, 'name': SearchDatastore_Task, 'duration_secs': 0.01172} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.863038] env[63088]: DEBUG oslo_concurrency.lockutils [None req-8044d1f3-cb0e-48b3-830d-fbfd6377d70c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1062.863424] env[63088]: DEBUG oslo_concurrency.lockutils [None req-8044d1f3-cb0e-48b3-830d-fbfd6377d70c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1062.865680] env[63088]: DEBUG nova.compute.manager [req-ba8cd0a1-c078-44fc-a869-56564890c74d req-4439ba65-a7cc-410a-bd68-417a11e98c8f service nova] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Received event network-vif-plugged-eb4d6233-2d9d-4756-8b9d-084bdfaeb3f7 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1062.866169] env[63088]: DEBUG oslo_concurrency.lockutils [req-ba8cd0a1-c078-44fc-a869-56564890c74d req-4439ba65-a7cc-410a-bd68-417a11e98c8f service nova] Acquiring lock "3fa11db4-2f70-495e-a562-ac5f06153847-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1062.866478] env[63088]: DEBUG oslo_concurrency.lockutils [req-ba8cd0a1-c078-44fc-a869-56564890c74d req-4439ba65-a7cc-410a-bd68-417a11e98c8f service nova] Lock "3fa11db4-2f70-495e-a562-ac5f06153847-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1062.866747] env[63088]: DEBUG oslo_concurrency.lockutils [req-ba8cd0a1-c078-44fc-a869-56564890c74d req-4439ba65-a7cc-410a-bd68-417a11e98c8f service nova] Lock "3fa11db4-2f70-495e-a562-ac5f06153847-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1062.867016] env[63088]: DEBUG nova.compute.manager [req-ba8cd0a1-c078-44fc-a869-56564890c74d req-4439ba65-a7cc-410a-bd68-417a11e98c8f service nova] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] No waiting events found dispatching network-vif-plugged-eb4d6233-2d9d-4756-8b9d-084bdfaeb3f7 {{(pid=63088) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1062.867276] env[63088]: WARNING nova.compute.manager [req-ba8cd0a1-c078-44fc-a869-56564890c74d req-4439ba65-a7cc-410a-bd68-417a11e98c8f service nova] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Received unexpected event network-vif-plugged-eb4d6233-2d9d-4756-8b9d-084bdfaeb3f7 for instance with vm_state building and task_state spawning. [ 1062.950375] env[63088]: DEBUG nova.network.neutron [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Successfully updated port: eb4d6233-2d9d-4756-8b9d-084bdfaeb3f7 {{(pid=63088) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1063.280756] env[63088]: DEBUG oslo_concurrency.lockutils [req-f215eb93-fa6e-4d54-ba13-e230bb080897 req-7616d55c-1c7a-423d-a689-67db4465b0e1 service nova] Releasing lock "refresh_cache-b25b5883-20c0-4cc9-8cc1-2c22a0f5be81" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1063.452914] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Acquiring lock "refresh_cache-3fa11db4-2f70-495e-a562-ac5f06153847" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1063.453077] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Acquired lock "refresh_cache-3fa11db4-2f70-495e-a562-ac5f06153847" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1063.453219] env[63088]: DEBUG nova.network.neutron [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1063.457427] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5f12d59-acbe-4e6e-a7f1-4150aef52567 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.466624] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1bf3e7b-b555-4187-9c77-6edfd39ae27d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.499873] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4feed88-31fe-4a86-b8d0-7d869efe1cfc {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.509021] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79b26dcd-ce07-4180-a6b6-2f96ad3c252f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.523945] env[63088]: DEBUG nova.compute.provider_tree [None req-8044d1f3-cb0e-48b3-830d-fbfd6377d70c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1063.986367] env[63088]: DEBUG nova.network.neutron [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1064.026870] env[63088]: DEBUG nova.scheduler.client.report [None req-8044d1f3-cb0e-48b3-830d-fbfd6377d70c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1064.155513] env[63088]: DEBUG nova.network.neutron [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Updating instance_info_cache with network_info: [{"id": "eb4d6233-2d9d-4756-8b9d-084bdfaeb3f7", "address": "fa:16:3e:c2:db:1d", "network": {"id": "cc2c0304-8906-443a-a76e-e7a32158346f", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1867758042-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b18ddddf2314d83addf550b8cb91977", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0f096917-a0cf-4add-a9d2-23ca1c723b3b", "external-id": "nsx-vlan-transportzone-894", "segmentation_id": 894, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeb4d6233-2d", "ovs_interfaceid": "eb4d6233-2d9d-4756-8b9d-084bdfaeb3f7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1064.658437] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Releasing lock "refresh_cache-3fa11db4-2f70-495e-a562-ac5f06153847" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1064.658806] env[63088]: DEBUG nova.compute.manager [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Instance network_info: |[{"id": "eb4d6233-2d9d-4756-8b9d-084bdfaeb3f7", "address": "fa:16:3e:c2:db:1d", "network": {"id": "cc2c0304-8906-443a-a76e-e7a32158346f", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1867758042-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b18ddddf2314d83addf550b8cb91977", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0f096917-a0cf-4add-a9d2-23ca1c723b3b", "external-id": "nsx-vlan-transportzone-894", "segmentation_id": 894, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeb4d6233-2d", "ovs_interfaceid": "eb4d6233-2d9d-4756-8b9d-084bdfaeb3f7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1064.659326] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c2:db:1d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0f096917-a0cf-4add-a9d2-23ca1c723b3b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'eb4d6233-2d9d-4756-8b9d-084bdfaeb3f7', 'vif_model': 'vmxnet3'}] {{(pid=63088) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1064.667267] env[63088]: DEBUG oslo.service.loopingcall [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1064.667520] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Creating VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1064.667761] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-df4c42b4-c15b-4513-a461-c1d4f3905439 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.687897] env[63088]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1064.687897] env[63088]: value = "task-1285512" [ 1064.687897] env[63088]: _type = "Task" [ 1064.687897] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.695624] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285512, 'name': CreateVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.887298] env[63088]: DEBUG nova.compute.manager [req-765b8905-757e-4042-bc64-0287cda002e2 req-ed91f05b-c526-4aeb-bc07-9caad7b48016 service nova] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Received event network-changed-eb4d6233-2d9d-4756-8b9d-084bdfaeb3f7 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1064.887505] env[63088]: DEBUG nova.compute.manager [req-765b8905-757e-4042-bc64-0287cda002e2 req-ed91f05b-c526-4aeb-bc07-9caad7b48016 service nova] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Refreshing instance network info cache due to event network-changed-eb4d6233-2d9d-4756-8b9d-084bdfaeb3f7. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1064.887739] env[63088]: DEBUG oslo_concurrency.lockutils [req-765b8905-757e-4042-bc64-0287cda002e2 req-ed91f05b-c526-4aeb-bc07-9caad7b48016 service nova] Acquiring lock "refresh_cache-3fa11db4-2f70-495e-a562-ac5f06153847" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1064.887898] env[63088]: DEBUG oslo_concurrency.lockutils [req-765b8905-757e-4042-bc64-0287cda002e2 req-ed91f05b-c526-4aeb-bc07-9caad7b48016 service nova] Acquired lock "refresh_cache-3fa11db4-2f70-495e-a562-ac5f06153847" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1064.888071] env[63088]: DEBUG nova.network.neutron [req-765b8905-757e-4042-bc64-0287cda002e2 req-ed91f05b-c526-4aeb-bc07-9caad7b48016 service nova] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Refreshing network info cache for port eb4d6233-2d9d-4756-8b9d-084bdfaeb3f7 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1065.038436] env[63088]: DEBUG oslo_concurrency.lockutils [None req-8044d1f3-cb0e-48b3-830d-fbfd6377d70c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.175s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1065.198248] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285512, 'name': CreateVM_Task, 'duration_secs': 0.45379} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.198435] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Created VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1065.199099] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1065.199275] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1065.199613] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1065.199980] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-11618ac9-232b-4a9f-a578-85d7779f57ed {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.204629] env[63088]: DEBUG oslo_vmware.api [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Waiting for the task: (returnval){ [ 1065.204629] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52640a7a-46d1-8984-7a39-871eb56cf314" [ 1065.204629] env[63088]: _type = "Task" [ 1065.204629] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.212771] env[63088]: DEBUG oslo_vmware.api [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52640a7a-46d1-8984-7a39-871eb56cf314, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.582892] env[63088]: DEBUG nova.network.neutron [req-765b8905-757e-4042-bc64-0287cda002e2 req-ed91f05b-c526-4aeb-bc07-9caad7b48016 service nova] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Updated VIF entry in instance network info cache for port eb4d6233-2d9d-4756-8b9d-084bdfaeb3f7. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1065.583339] env[63088]: DEBUG nova.network.neutron [req-765b8905-757e-4042-bc64-0287cda002e2 req-ed91f05b-c526-4aeb-bc07-9caad7b48016 service nova] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Updating instance_info_cache with network_info: [{"id": "eb4d6233-2d9d-4756-8b9d-084bdfaeb3f7", "address": "fa:16:3e:c2:db:1d", "network": {"id": "cc2c0304-8906-443a-a76e-e7a32158346f", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1867758042-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b18ddddf2314d83addf550b8cb91977", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0f096917-a0cf-4add-a9d2-23ca1c723b3b", "external-id": "nsx-vlan-transportzone-894", "segmentation_id": 894, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeb4d6233-2d", "ovs_interfaceid": "eb4d6233-2d9d-4756-8b9d-084bdfaeb3f7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1065.592920] env[63088]: INFO nova.scheduler.client.report [None req-8044d1f3-cb0e-48b3-830d-fbfd6377d70c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Deleted allocation for migration 1873436d-a9d6-4d6c-80b5-d6b5decb0c77 [ 1065.716343] env[63088]: DEBUG oslo_vmware.api [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52640a7a-46d1-8984-7a39-871eb56cf314, 'name': SearchDatastore_Task, 'duration_secs': 0.012262} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.716676] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1065.716983] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Processing image 1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1065.717288] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1065.717525] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1065.717744] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1065.718105] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-45eeb369-77d8-4b6c-bf81-d0b079f75f3a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.726586] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1065.726797] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1065.727730] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-af9e58e7-950c-4f1f-907b-424ccca28763 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.733422] env[63088]: DEBUG oslo_vmware.api [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Waiting for the task: (returnval){ [ 1065.733422] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52d6bc06-9922-35bc-a715-b9c52383363c" [ 1065.733422] env[63088]: _type = "Task" [ 1065.733422] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.740840] env[63088]: DEBUG oslo_vmware.api [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52d6bc06-9922-35bc-a715-b9c52383363c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.086080] env[63088]: DEBUG oslo_concurrency.lockutils [req-765b8905-757e-4042-bc64-0287cda002e2 req-ed91f05b-c526-4aeb-bc07-9caad7b48016 service nova] Releasing lock "refresh_cache-3fa11db4-2f70-495e-a562-ac5f06153847" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1066.098552] env[63088]: DEBUG oslo_concurrency.lockutils [None req-8044d1f3-cb0e-48b3-830d-fbfd6377d70c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "d94e57a3-743b-4bfe-b208-80f5748f8654" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.741s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1066.244644] env[63088]: DEBUG oslo_vmware.api [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52d6bc06-9922-35bc-a715-b9c52383363c, 'name': SearchDatastore_Task, 'duration_secs': 0.012473} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.245513] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d7a89dc7-fae0-4d8a-8a89-2a25a5e7829e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.251494] env[63088]: DEBUG oslo_vmware.api [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Waiting for the task: (returnval){ [ 1066.251494] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52eedb86-72e2-e0b9-0b64-448bde62ded5" [ 1066.251494] env[63088]: _type = "Task" [ 1066.251494] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.259252] env[63088]: DEBUG oslo_vmware.api [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52eedb86-72e2-e0b9-0b64-448bde62ded5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.687051] env[63088]: DEBUG oslo_vmware.rw_handles [None req-72711027-105f-439f-a56d-b2a885190dae tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52083c3a-9df7-e61e-1da8-93e0f94c74d7/disk-0.vmdk. {{(pid=63088) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1066.688054] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e281b6e4-b761-44af-985a-13ce64bb1093 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.694394] env[63088]: DEBUG oslo_vmware.rw_handles [None req-72711027-105f-439f-a56d-b2a885190dae tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52083c3a-9df7-e61e-1da8-93e0f94c74d7/disk-0.vmdk is in state: ready. {{(pid=63088) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1066.694555] env[63088]: ERROR oslo_vmware.rw_handles [None req-72711027-105f-439f-a56d-b2a885190dae tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52083c3a-9df7-e61e-1da8-93e0f94c74d7/disk-0.vmdk due to incomplete transfer. [ 1066.694766] env[63088]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-57043d78-2fe2-4505-a8d1-662dd7862fae {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.701036] env[63088]: DEBUG oslo_vmware.rw_handles [None req-72711027-105f-439f-a56d-b2a885190dae tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52083c3a-9df7-e61e-1da8-93e0f94c74d7/disk-0.vmdk. {{(pid=63088) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1066.701238] env[63088]: DEBUG nova.virt.vmwareapi.images [None req-72711027-105f-439f-a56d-b2a885190dae tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Uploaded image 7862f3ca-7713-4ff8-8498-b50e9658c73f to the Glance image server {{(pid=63088) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1066.703492] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-72711027-105f-439f-a56d-b2a885190dae tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Destroying the VM {{(pid=63088) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1066.703719] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-d2e45178-6248-4d07-8e1b-cd894c03a462 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.709614] env[63088]: DEBUG oslo_vmware.api [None req-72711027-105f-439f-a56d-b2a885190dae tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1066.709614] env[63088]: value = "task-1285513" [ 1066.709614] env[63088]: _type = "Task" [ 1066.709614] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.717288] env[63088]: DEBUG oslo_vmware.api [None req-72711027-105f-439f-a56d-b2a885190dae tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285513, 'name': Destroy_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.761922] env[63088]: DEBUG oslo_vmware.api [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52eedb86-72e2-e0b9-0b64-448bde62ded5, 'name': SearchDatastore_Task, 'duration_secs': 0.014481} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.762187] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1066.762450] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] 3fa11db4-2f70-495e-a562-ac5f06153847/3fa11db4-2f70-495e-a562-ac5f06153847.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1066.762702] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f9c02020-3ba0-4239-8f3c-e080bc0fe820 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.769086] env[63088]: DEBUG oslo_vmware.api [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Waiting for the task: (returnval){ [ 1066.769086] env[63088]: value = "task-1285514" [ 1066.769086] env[63088]: _type = "Task" [ 1066.769086] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.776216] env[63088]: DEBUG oslo_vmware.api [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285514, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.220259] env[63088]: DEBUG oslo_vmware.api [None req-72711027-105f-439f-a56d-b2a885190dae tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285513, 'name': Destroy_Task} progress is 33%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.279505] env[63088]: DEBUG oslo_vmware.api [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285514, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.381045] env[63088]: DEBUG oslo_concurrency.lockutils [None req-8044d1f3-cb0e-48b3-830d-fbfd6377d70c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquiring lock "d94e57a3-743b-4bfe-b208-80f5748f8654" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1067.381045] env[63088]: DEBUG oslo_concurrency.lockutils [None req-8044d1f3-cb0e-48b3-830d-fbfd6377d70c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "d94e57a3-743b-4bfe-b208-80f5748f8654" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1067.381045] env[63088]: DEBUG oslo_concurrency.lockutils [None req-8044d1f3-cb0e-48b3-830d-fbfd6377d70c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquiring lock "d94e57a3-743b-4bfe-b208-80f5748f8654-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1067.381045] env[63088]: DEBUG oslo_concurrency.lockutils [None req-8044d1f3-cb0e-48b3-830d-fbfd6377d70c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "d94e57a3-743b-4bfe-b208-80f5748f8654-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1067.381423] env[63088]: DEBUG oslo_concurrency.lockutils [None req-8044d1f3-cb0e-48b3-830d-fbfd6377d70c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "d94e57a3-743b-4bfe-b208-80f5748f8654-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1067.383417] env[63088]: INFO nova.compute.manager [None req-8044d1f3-cb0e-48b3-830d-fbfd6377d70c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Terminating instance [ 1067.385168] env[63088]: DEBUG nova.compute.manager [None req-8044d1f3-cb0e-48b3-830d-fbfd6377d70c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1067.385372] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-8044d1f3-cb0e-48b3-830d-fbfd6377d70c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1067.386235] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-807b21a9-f70c-4599-b76d-df5d3065230e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.393878] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-8044d1f3-cb0e-48b3-830d-fbfd6377d70c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1067.394126] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b5ee83b2-f310-4dba-90f9-250e7a56d13e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.400660] env[63088]: DEBUG oslo_vmware.api [None req-8044d1f3-cb0e-48b3-830d-fbfd6377d70c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 1067.400660] env[63088]: value = "task-1285515" [ 1067.400660] env[63088]: _type = "Task" [ 1067.400660] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.408692] env[63088]: DEBUG oslo_vmware.api [None req-8044d1f3-cb0e-48b3-830d-fbfd6377d70c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285515, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.720156] env[63088]: DEBUG oslo_vmware.api [None req-72711027-105f-439f-a56d-b2a885190dae tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285513, 'name': Destroy_Task, 'duration_secs': 0.716871} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.720434] env[63088]: INFO nova.virt.vmwareapi.vm_util [None req-72711027-105f-439f-a56d-b2a885190dae tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Destroyed the VM [ 1067.720678] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-72711027-105f-439f-a56d-b2a885190dae tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Deleting Snapshot of the VM instance {{(pid=63088) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1067.720934] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-f69b9ab9-eb09-48f3-b146-ad8086aa0253 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.727368] env[63088]: DEBUG oslo_vmware.api [None req-72711027-105f-439f-a56d-b2a885190dae tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1067.727368] env[63088]: value = "task-1285516" [ 1067.727368] env[63088]: _type = "Task" [ 1067.727368] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.736424] env[63088]: DEBUG oslo_vmware.api [None req-72711027-105f-439f-a56d-b2a885190dae tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285516, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.779173] env[63088]: DEBUG oslo_vmware.api [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285514, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.635688} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.779435] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] 3fa11db4-2f70-495e-a562-ac5f06153847/3fa11db4-2f70-495e-a562-ac5f06153847.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1067.779647] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Extending root virtual disk to 1048576 {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1067.779922] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-47e74437-d981-45a5-9fdc-db3ec303be45 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.786082] env[63088]: DEBUG oslo_vmware.api [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Waiting for the task: (returnval){ [ 1067.786082] env[63088]: value = "task-1285517" [ 1067.786082] env[63088]: _type = "Task" [ 1067.786082] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.793802] env[63088]: DEBUG oslo_vmware.api [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285517, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.910233] env[63088]: DEBUG oslo_vmware.api [None req-8044d1f3-cb0e-48b3-830d-fbfd6377d70c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285515, 'name': PowerOffVM_Task, 'duration_secs': 0.187863} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.910515] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-8044d1f3-cb0e-48b3-830d-fbfd6377d70c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1067.911119] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-8044d1f3-cb0e-48b3-830d-fbfd6377d70c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Unregistering the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1067.911119] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-328f65b5-32b7-40ff-bc49-eac4924f42ab {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.985020] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-8044d1f3-cb0e-48b3-830d-fbfd6377d70c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Unregistered the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1067.985270] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-8044d1f3-cb0e-48b3-830d-fbfd6377d70c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Deleting contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1067.985462] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-8044d1f3-cb0e-48b3-830d-fbfd6377d70c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Deleting the datastore file [datastore1] d94e57a3-743b-4bfe-b208-80f5748f8654 {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1067.985732] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a80f7fd9-4673-4051-8f50-410bd87c9b6e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.992473] env[63088]: DEBUG oslo_vmware.api [None req-8044d1f3-cb0e-48b3-830d-fbfd6377d70c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for the task: (returnval){ [ 1067.992473] env[63088]: value = "task-1285519" [ 1067.992473] env[63088]: _type = "Task" [ 1067.992473] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.999447] env[63088]: DEBUG oslo_vmware.api [None req-8044d1f3-cb0e-48b3-830d-fbfd6377d70c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285519, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.237128] env[63088]: DEBUG oslo_vmware.api [None req-72711027-105f-439f-a56d-b2a885190dae tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285516, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.295842] env[63088]: DEBUG oslo_vmware.api [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285517, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065965} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.296085] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Extended root virtual disk {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1068.296859] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fb6f765-6c07-4f6a-8d24-7c15163ad2c6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.318364] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Reconfiguring VM instance instance-00000068 to attach disk [datastore1] 3fa11db4-2f70-495e-a562-ac5f06153847/3fa11db4-2f70-495e-a562-ac5f06153847.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1068.318626] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c407aab8-28c1-4fd5-9685-7c5cf0c49a25 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.337719] env[63088]: DEBUG oslo_vmware.api [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Waiting for the task: (returnval){ [ 1068.337719] env[63088]: value = "task-1285520" [ 1068.337719] env[63088]: _type = "Task" [ 1068.337719] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.345485] env[63088]: DEBUG oslo_vmware.api [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285520, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.502141] env[63088]: DEBUG oslo_vmware.api [None req-8044d1f3-cb0e-48b3-830d-fbfd6377d70c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Task: {'id': task-1285519, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.158256} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.502400] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-8044d1f3-cb0e-48b3-830d-fbfd6377d70c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1068.502592] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-8044d1f3-cb0e-48b3-830d-fbfd6377d70c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Deleted contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1068.502778] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-8044d1f3-cb0e-48b3-830d-fbfd6377d70c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1068.502959] env[63088]: INFO nova.compute.manager [None req-8044d1f3-cb0e-48b3-830d-fbfd6377d70c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1068.503232] env[63088]: DEBUG oslo.service.loopingcall [None req-8044d1f3-cb0e-48b3-830d-fbfd6377d70c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1068.503637] env[63088]: DEBUG nova.compute.manager [-] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1068.503749] env[63088]: DEBUG nova.network.neutron [-] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1068.738707] env[63088]: DEBUG oslo_vmware.api [None req-72711027-105f-439f-a56d-b2a885190dae tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285516, 'name': RemoveSnapshot_Task, 'duration_secs': 0.526958} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.739171] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-72711027-105f-439f-a56d-b2a885190dae tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Deleted Snapshot of the VM instance {{(pid=63088) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1068.739479] env[63088]: INFO nova.compute.manager [None req-72711027-105f-439f-a56d-b2a885190dae tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Took 12.76 seconds to snapshot the instance on the hypervisor. [ 1068.833113] env[63088]: DEBUG nova.compute.manager [req-affbe89b-1e8c-4e8e-b0d5-6a7933139122 req-73080ee1-3d9e-4a00-8afd-55910c073289 service nova] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Received event network-vif-deleted-024c5e7f-23a2-46df-83c1-5195ff662b68 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1068.833322] env[63088]: INFO nova.compute.manager [req-affbe89b-1e8c-4e8e-b0d5-6a7933139122 req-73080ee1-3d9e-4a00-8afd-55910c073289 service nova] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Neutron deleted interface 024c5e7f-23a2-46df-83c1-5195ff662b68; detaching it from the instance and deleting it from the info cache [ 1068.833501] env[63088]: DEBUG nova.network.neutron [req-affbe89b-1e8c-4e8e-b0d5-6a7933139122 req-73080ee1-3d9e-4a00-8afd-55910c073289 service nova] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1068.848815] env[63088]: DEBUG oslo_vmware.api [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285520, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.294652] env[63088]: DEBUG nova.compute.manager [None req-72711027-105f-439f-a56d-b2a885190dae tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Found 2 images (rotation: 2) {{(pid=63088) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 1069.309914] env[63088]: DEBUG nova.network.neutron [-] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1069.335862] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4a370a8d-2f3c-4347-927d-7a90ffcc5907 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.346131] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09f75302-d01c-4a32-af23-03eac89cea35 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.359074] env[63088]: DEBUG oslo_vmware.api [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285520, 'name': ReconfigVM_Task, 'duration_secs': 0.732148} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.359694] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Reconfigured VM instance instance-00000068 to attach disk [datastore1] 3fa11db4-2f70-495e-a562-ac5f06153847/3fa11db4-2f70-495e-a562-ac5f06153847.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1069.360346] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a5493974-98eb-43e1-a1c8-5c4dadfe0be2 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.373507] env[63088]: DEBUG nova.compute.manager [req-affbe89b-1e8c-4e8e-b0d5-6a7933139122 req-73080ee1-3d9e-4a00-8afd-55910c073289 service nova] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Detach interface failed, port_id=024c5e7f-23a2-46df-83c1-5195ff662b68, reason: Instance d94e57a3-743b-4bfe-b208-80f5748f8654 could not be found. {{(pid=63088) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1069.374786] env[63088]: DEBUG oslo_vmware.api [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Waiting for the task: (returnval){ [ 1069.374786] env[63088]: value = "task-1285521" [ 1069.374786] env[63088]: _type = "Task" [ 1069.374786] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.384465] env[63088]: DEBUG oslo_vmware.api [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285521, 'name': Rename_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.753534] env[63088]: DEBUG nova.compute.manager [None req-8563ac7d-070a-4547-bb9e-708353533728 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1069.754492] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83e2e503-68f4-48b5-8a89-98963f250041 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.812142] env[63088]: INFO nova.compute.manager [-] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Took 1.31 seconds to deallocate network for instance. [ 1069.884561] env[63088]: DEBUG oslo_vmware.api [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285521, 'name': Rename_Task, 'duration_secs': 0.135313} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.884806] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1069.885118] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4d9bc49e-6455-486e-9529-45aca4e4f095 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.891782] env[63088]: DEBUG oslo_vmware.api [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Waiting for the task: (returnval){ [ 1069.891782] env[63088]: value = "task-1285522" [ 1069.891782] env[63088]: _type = "Task" [ 1069.891782] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.899020] env[63088]: DEBUG oslo_vmware.api [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285522, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.266310] env[63088]: INFO nova.compute.manager [None req-8563ac7d-070a-4547-bb9e-708353533728 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] instance snapshotting [ 1070.266980] env[63088]: DEBUG nova.objects.instance [None req-8563ac7d-070a-4547-bb9e-708353533728 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lazy-loading 'flavor' on Instance uuid 2b87a9f1-5f10-43c2-8bc9-6d560ec88015 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1070.319986] env[63088]: DEBUG oslo_concurrency.lockutils [None req-8044d1f3-cb0e-48b3-830d-fbfd6377d70c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1070.320383] env[63088]: DEBUG oslo_concurrency.lockutils [None req-8044d1f3-cb0e-48b3-830d-fbfd6377d70c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1070.320449] env[63088]: DEBUG oslo_concurrency.lockutils [None req-8044d1f3-cb0e-48b3-830d-fbfd6377d70c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1070.343642] env[63088]: INFO nova.scheduler.client.report [None req-8044d1f3-cb0e-48b3-830d-fbfd6377d70c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Deleted allocations for instance d94e57a3-743b-4bfe-b208-80f5748f8654 [ 1070.404555] env[63088]: DEBUG oslo_vmware.api [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285522, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.774357] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c37ad7bc-1252-4bd2-85cb-ced334da8a32 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.796115] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6afe625-5e35-45a3-a8e6-023e6e9d6b07 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.851896] env[63088]: DEBUG oslo_concurrency.lockutils [None req-8044d1f3-cb0e-48b3-830d-fbfd6377d70c tempest-DeleteServersTestJSON-929036285 tempest-DeleteServersTestJSON-929036285-project-member] Lock "d94e57a3-743b-4bfe-b208-80f5748f8654" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.471s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1070.902863] env[63088]: DEBUG oslo_vmware.api [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285522, 'name': PowerOnVM_Task, 'duration_secs': 0.67471} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.903553] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1070.903830] env[63088]: INFO nova.compute.manager [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Took 8.54 seconds to spawn the instance on the hypervisor. [ 1070.904035] env[63088]: DEBUG nova.compute.manager [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1070.905516] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-829f6034-3bc2-49a8-bac2-13e28236f685 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.307649] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-8563ac7d-070a-4547-bb9e-708353533728 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Creating Snapshot of the VM instance {{(pid=63088) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1071.308354] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-67b3aa9b-2cc7-4b27-8f18-6438df6c73e4 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.316478] env[63088]: DEBUG oslo_vmware.api [None req-8563ac7d-070a-4547-bb9e-708353533728 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1071.316478] env[63088]: value = "task-1285523" [ 1071.316478] env[63088]: _type = "Task" [ 1071.316478] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.324021] env[63088]: DEBUG oslo_vmware.api [None req-8563ac7d-070a-4547-bb9e-708353533728 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285523, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.425831] env[63088]: INFO nova.compute.manager [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Took 13.28 seconds to build instance. [ 1071.826530] env[63088]: DEBUG oslo_vmware.api [None req-8563ac7d-070a-4547-bb9e-708353533728 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285523, 'name': CreateSnapshot_Task, 'duration_secs': 0.498031} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.826795] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-8563ac7d-070a-4547-bb9e-708353533728 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Created Snapshot of the VM instance {{(pid=63088) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1071.827538] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e88b69d9-60ec-47b5-9984-f6c5134f24c3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.927477] env[63088]: DEBUG oslo_concurrency.lockutils [None req-3f83915f-bd97-453b-af18-26f6b6e6c36c tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Lock "3fa11db4-2f70-495e-a562-ac5f06153847" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.789s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1072.344492] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-8563ac7d-070a-4547-bb9e-708353533728 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Creating linked-clone VM from snapshot {{(pid=63088) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1072.344828] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-45fe8157-dacb-4d4b-b5d1-bfa59ce2f7a9 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.354457] env[63088]: DEBUG oslo_vmware.api [None req-8563ac7d-070a-4547-bb9e-708353533728 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1072.354457] env[63088]: value = "task-1285525" [ 1072.354457] env[63088]: _type = "Task" [ 1072.354457] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.362297] env[63088]: DEBUG oslo_vmware.api [None req-8563ac7d-070a-4547-bb9e-708353533728 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285525, 'name': CloneVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.552239] env[63088]: DEBUG nova.compute.manager [req-fb454db6-3bd6-44cb-8b05-b0e697d0de3b req-2a95f543-b614-41e8-83c3-d156ca6777da service nova] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Received event network-changed-eb4d6233-2d9d-4756-8b9d-084bdfaeb3f7 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1072.552239] env[63088]: DEBUG nova.compute.manager [req-fb454db6-3bd6-44cb-8b05-b0e697d0de3b req-2a95f543-b614-41e8-83c3-d156ca6777da service nova] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Refreshing instance network info cache due to event network-changed-eb4d6233-2d9d-4756-8b9d-084bdfaeb3f7. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1072.552512] env[63088]: DEBUG oslo_concurrency.lockutils [req-fb454db6-3bd6-44cb-8b05-b0e697d0de3b req-2a95f543-b614-41e8-83c3-d156ca6777da service nova] Acquiring lock "refresh_cache-3fa11db4-2f70-495e-a562-ac5f06153847" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1072.552693] env[63088]: DEBUG oslo_concurrency.lockutils [req-fb454db6-3bd6-44cb-8b05-b0e697d0de3b req-2a95f543-b614-41e8-83c3-d156ca6777da service nova] Acquired lock "refresh_cache-3fa11db4-2f70-495e-a562-ac5f06153847" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1072.552876] env[63088]: DEBUG nova.network.neutron [req-fb454db6-3bd6-44cb-8b05-b0e697d0de3b req-2a95f543-b614-41e8-83c3-d156ca6777da service nova] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Refreshing network info cache for port eb4d6233-2d9d-4756-8b9d-084bdfaeb3f7 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1072.864814] env[63088]: DEBUG oslo_vmware.api [None req-8563ac7d-070a-4547-bb9e-708353533728 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285525, 'name': CloneVM_Task} progress is 94%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.270605] env[63088]: DEBUG nova.network.neutron [req-fb454db6-3bd6-44cb-8b05-b0e697d0de3b req-2a95f543-b614-41e8-83c3-d156ca6777da service nova] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Updated VIF entry in instance network info cache for port eb4d6233-2d9d-4756-8b9d-084bdfaeb3f7. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1073.270990] env[63088]: DEBUG nova.network.neutron [req-fb454db6-3bd6-44cb-8b05-b0e697d0de3b req-2a95f543-b614-41e8-83c3-d156ca6777da service nova] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Updating instance_info_cache with network_info: [{"id": "eb4d6233-2d9d-4756-8b9d-084bdfaeb3f7", "address": "fa:16:3e:c2:db:1d", "network": {"id": "cc2c0304-8906-443a-a76e-e7a32158346f", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1867758042-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.195", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b18ddddf2314d83addf550b8cb91977", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0f096917-a0cf-4add-a9d2-23ca1c723b3b", "external-id": "nsx-vlan-transportzone-894", "segmentation_id": 894, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeb4d6233-2d", "ovs_interfaceid": "eb4d6233-2d9d-4756-8b9d-084bdfaeb3f7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1073.364367] env[63088]: DEBUG oslo_vmware.api [None req-8563ac7d-070a-4547-bb9e-708353533728 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285525, 'name': CloneVM_Task, 'duration_secs': 0.956426} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.364731] env[63088]: INFO nova.virt.vmwareapi.vmops [None req-8563ac7d-070a-4547-bb9e-708353533728 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Created linked-clone VM from snapshot [ 1073.365360] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6af465f-2b97-47b8-b467-6d3b40c29a97 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.372319] env[63088]: DEBUG nova.virt.vmwareapi.images [None req-8563ac7d-070a-4547-bb9e-708353533728 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Uploading image efd7e2d8-5c54-435f-9efa-6982f84bf424 {{(pid=63088) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1073.393518] env[63088]: DEBUG oslo_vmware.rw_handles [None req-8563ac7d-070a-4547-bb9e-708353533728 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1073.393518] env[63088]: value = "vm-275961" [ 1073.393518] env[63088]: _type = "VirtualMachine" [ 1073.393518] env[63088]: }. {{(pid=63088) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1073.393766] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-ccbe4f6e-28a5-444d-b880-ca4f7f551ea8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.400285] env[63088]: DEBUG oslo_vmware.rw_handles [None req-8563ac7d-070a-4547-bb9e-708353533728 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lease: (returnval){ [ 1073.400285] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]5204f43e-8584-b2f9-cc78-362c9c1c06b6" [ 1073.400285] env[63088]: _type = "HttpNfcLease" [ 1073.400285] env[63088]: } obtained for exporting VM: (result){ [ 1073.400285] env[63088]: value = "vm-275961" [ 1073.400285] env[63088]: _type = "VirtualMachine" [ 1073.400285] env[63088]: }. {{(pid=63088) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1073.400495] env[63088]: DEBUG oslo_vmware.api [None req-8563ac7d-070a-4547-bb9e-708353533728 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the lease: (returnval){ [ 1073.400495] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]5204f43e-8584-b2f9-cc78-362c9c1c06b6" [ 1073.400495] env[63088]: _type = "HttpNfcLease" [ 1073.400495] env[63088]: } to be ready. {{(pid=63088) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1073.406209] env[63088]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1073.406209] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]5204f43e-8584-b2f9-cc78-362c9c1c06b6" [ 1073.406209] env[63088]: _type = "HttpNfcLease" [ 1073.406209] env[63088]: } is initializing. {{(pid=63088) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1073.774134] env[63088]: DEBUG oslo_concurrency.lockutils [req-fb454db6-3bd6-44cb-8b05-b0e697d0de3b req-2a95f543-b614-41e8-83c3-d156ca6777da service nova] Releasing lock "refresh_cache-3fa11db4-2f70-495e-a562-ac5f06153847" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1073.908761] env[63088]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1073.908761] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]5204f43e-8584-b2f9-cc78-362c9c1c06b6" [ 1073.908761] env[63088]: _type = "HttpNfcLease" [ 1073.908761] env[63088]: } is ready. {{(pid=63088) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1073.908982] env[63088]: DEBUG oslo_vmware.rw_handles [None req-8563ac7d-070a-4547-bb9e-708353533728 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1073.908982] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]5204f43e-8584-b2f9-cc78-362c9c1c06b6" [ 1073.908982] env[63088]: _type = "HttpNfcLease" [ 1073.908982] env[63088]: }. {{(pid=63088) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1073.909699] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cdd7f29-10cd-4384-84d1-8bef38a3419f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.916537] env[63088]: DEBUG oslo_vmware.rw_handles [None req-8563ac7d-070a-4547-bb9e-708353533728 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52f016e5-f1ec-fe75-2196-42b2319eafea/disk-0.vmdk from lease info. {{(pid=63088) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1073.916710] env[63088]: DEBUG oslo_vmware.rw_handles [None req-8563ac7d-070a-4547-bb9e-708353533728 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52f016e5-f1ec-fe75-2196-42b2319eafea/disk-0.vmdk for reading. {{(pid=63088) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1074.002715] env[63088]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-ae04cacd-9c27-4dca-8fdb-d9dc296cf2a7 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.605669] env[63088]: DEBUG oslo_vmware.rw_handles [None req-8563ac7d-070a-4547-bb9e-708353533728 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52f016e5-f1ec-fe75-2196-42b2319eafea/disk-0.vmdk. {{(pid=63088) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1081.606722] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5e95cc2-62d2-49d8-8a60-0446b6793a02 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.612968] env[63088]: DEBUG oslo_vmware.rw_handles [None req-8563ac7d-070a-4547-bb9e-708353533728 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52f016e5-f1ec-fe75-2196-42b2319eafea/disk-0.vmdk is in state: ready. {{(pid=63088) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1081.613152] env[63088]: ERROR oslo_vmware.rw_handles [None req-8563ac7d-070a-4547-bb9e-708353533728 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52f016e5-f1ec-fe75-2196-42b2319eafea/disk-0.vmdk due to incomplete transfer. [ 1081.613369] env[63088]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-c4a41d63-588f-4fbd-9c52-1c9101fa65e1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.619993] env[63088]: DEBUG oslo_vmware.rw_handles [None req-8563ac7d-070a-4547-bb9e-708353533728 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52f016e5-f1ec-fe75-2196-42b2319eafea/disk-0.vmdk. {{(pid=63088) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1081.620226] env[63088]: DEBUG nova.virt.vmwareapi.images [None req-8563ac7d-070a-4547-bb9e-708353533728 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Uploaded image efd7e2d8-5c54-435f-9efa-6982f84bf424 to the Glance image server {{(pid=63088) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1081.622533] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-8563ac7d-070a-4547-bb9e-708353533728 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Destroying the VM {{(pid=63088) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1081.622761] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-0c60cd62-892f-420e-b2b1-3994e5c8e268 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.628890] env[63088]: DEBUG oslo_vmware.api [None req-8563ac7d-070a-4547-bb9e-708353533728 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1081.628890] env[63088]: value = "task-1285527" [ 1081.628890] env[63088]: _type = "Task" [ 1081.628890] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.637430] env[63088]: DEBUG oslo_vmware.api [None req-8563ac7d-070a-4547-bb9e-708353533728 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285527, 'name': Destroy_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.138802] env[63088]: DEBUG oslo_vmware.api [None req-8563ac7d-070a-4547-bb9e-708353533728 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285527, 'name': Destroy_Task} progress is 100%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.639278] env[63088]: DEBUG oslo_vmware.api [None req-8563ac7d-070a-4547-bb9e-708353533728 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285527, 'name': Destroy_Task} progress is 100%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.140273] env[63088]: DEBUG oslo_vmware.api [None req-8563ac7d-070a-4547-bb9e-708353533728 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285527, 'name': Destroy_Task, 'duration_secs': 1.295808} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.140565] env[63088]: INFO nova.virt.vmwareapi.vm_util [None req-8563ac7d-070a-4547-bb9e-708353533728 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Destroyed the VM [ 1083.140799] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-8563ac7d-070a-4547-bb9e-708353533728 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Deleting Snapshot of the VM instance {{(pid=63088) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1083.141066] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-f1e3a052-1230-4b0b-9b45-a0df379cb214 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.147336] env[63088]: DEBUG oslo_vmware.api [None req-8563ac7d-070a-4547-bb9e-708353533728 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1083.147336] env[63088]: value = "task-1285528" [ 1083.147336] env[63088]: _type = "Task" [ 1083.147336] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.156664] env[63088]: DEBUG oslo_vmware.api [None req-8563ac7d-070a-4547-bb9e-708353533728 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285528, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.657541] env[63088]: DEBUG oslo_vmware.api [None req-8563ac7d-070a-4547-bb9e-708353533728 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285528, 'name': RemoveSnapshot_Task, 'duration_secs': 0.479303} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.657831] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-8563ac7d-070a-4547-bb9e-708353533728 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Deleted Snapshot of the VM instance {{(pid=63088) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1083.658074] env[63088]: INFO nova.compute.manager [None req-8563ac7d-070a-4547-bb9e-708353533728 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Took 12.88 seconds to snapshot the instance on the hypervisor. [ 1084.194501] env[63088]: DEBUG nova.compute.manager [None req-8563ac7d-070a-4547-bb9e-708353533728 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Found 2 images (rotation: 2) {{(pid=63088) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 1085.719521] env[63088]: DEBUG nova.compute.manager [None req-9e719bdd-5361-4621-a66f-918aaed14f4c tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1085.720508] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ca7d9c7-87e9-45f5-b2c8-6899f03638a9 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.230896] env[63088]: INFO nova.compute.manager [None req-9e719bdd-5361-4621-a66f-918aaed14f4c tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] instance snapshotting [ 1086.231625] env[63088]: DEBUG nova.objects.instance [None req-9e719bdd-5361-4621-a66f-918aaed14f4c tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lazy-loading 'flavor' on Instance uuid 2b87a9f1-5f10-43c2-8bc9-6d560ec88015 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1086.739594] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5de2ffe-6eaa-4be3-9ea9-a157f4a449df {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.759222] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54eca369-12e8-417b-87be-cf9c44f03fcf {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.269287] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-9e719bdd-5361-4621-a66f-918aaed14f4c tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Creating Snapshot of the VM instance {{(pid=63088) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1087.269597] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-201527b1-782c-4ef6-aba8-083848b5ac1f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.277620] env[63088]: DEBUG oslo_vmware.api [None req-9e719bdd-5361-4621-a66f-918aaed14f4c tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1087.277620] env[63088]: value = "task-1285529" [ 1087.277620] env[63088]: _type = "Task" [ 1087.277620] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.285355] env[63088]: DEBUG oslo_vmware.api [None req-9e719bdd-5361-4621-a66f-918aaed14f4c tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285529, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.787082] env[63088]: DEBUG oslo_vmware.api [None req-9e719bdd-5361-4621-a66f-918aaed14f4c tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285529, 'name': CreateSnapshot_Task, 'duration_secs': 0.396411} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.787450] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-9e719bdd-5361-4621-a66f-918aaed14f4c tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Created Snapshot of the VM instance {{(pid=63088) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1087.788036] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7487cc14-a205-4bfa-b4c4-8b7a1ce79b95 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.304917] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-9e719bdd-5361-4621-a66f-918aaed14f4c tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Creating linked-clone VM from snapshot {{(pid=63088) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1088.305234] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-3af47c74-ef0b-4f13-b0af-1b16e1f9faa7 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.314408] env[63088]: DEBUG oslo_vmware.api [None req-9e719bdd-5361-4621-a66f-918aaed14f4c tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1088.314408] env[63088]: value = "task-1285530" [ 1088.314408] env[63088]: _type = "Task" [ 1088.314408] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.321764] env[63088]: DEBUG oslo_vmware.api [None req-9e719bdd-5361-4621-a66f-918aaed14f4c tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285530, 'name': CloneVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.826860] env[63088]: DEBUG oslo_vmware.api [None req-9e719bdd-5361-4621-a66f-918aaed14f4c tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285530, 'name': CloneVM_Task} progress is 94%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.325990] env[63088]: DEBUG oslo_vmware.api [None req-9e719bdd-5361-4621-a66f-918aaed14f4c tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285530, 'name': CloneVM_Task} progress is 95%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.826301] env[63088]: DEBUG oslo_vmware.api [None req-9e719bdd-5361-4621-a66f-918aaed14f4c tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285530, 'name': CloneVM_Task, 'duration_secs': 1.288748} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.826565] env[63088]: INFO nova.virt.vmwareapi.vmops [None req-9e719bdd-5361-4621-a66f-918aaed14f4c tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Created linked-clone VM from snapshot [ 1089.827314] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-961c063a-d42d-4371-be0c-e3e8418309e5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.834284] env[63088]: DEBUG nova.virt.vmwareapi.images [None req-9e719bdd-5361-4621-a66f-918aaed14f4c tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Uploading image ecd13d17-7e92-4cbe-805f-9c202f437a2f {{(pid=63088) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1089.859830] env[63088]: DEBUG oslo_vmware.rw_handles [None req-9e719bdd-5361-4621-a66f-918aaed14f4c tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1089.859830] env[63088]: value = "vm-275963" [ 1089.859830] env[63088]: _type = "VirtualMachine" [ 1089.859830] env[63088]: }. {{(pid=63088) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1089.860123] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-a2f256c5-7597-45dd-8153-deea24eab3fb {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.867154] env[63088]: DEBUG oslo_vmware.rw_handles [None req-9e719bdd-5361-4621-a66f-918aaed14f4c tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lease: (returnval){ [ 1089.867154] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52fb97b9-8eca-b3e6-1723-ae511c9202ed" [ 1089.867154] env[63088]: _type = "HttpNfcLease" [ 1089.867154] env[63088]: } obtained for exporting VM: (result){ [ 1089.867154] env[63088]: value = "vm-275963" [ 1089.867154] env[63088]: _type = "VirtualMachine" [ 1089.867154] env[63088]: }. {{(pid=63088) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1089.867384] env[63088]: DEBUG oslo_vmware.api [None req-9e719bdd-5361-4621-a66f-918aaed14f4c tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the lease: (returnval){ [ 1089.867384] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52fb97b9-8eca-b3e6-1723-ae511c9202ed" [ 1089.867384] env[63088]: _type = "HttpNfcLease" [ 1089.867384] env[63088]: } to be ready. {{(pid=63088) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1089.873377] env[63088]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1089.873377] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52fb97b9-8eca-b3e6-1723-ae511c9202ed" [ 1089.873377] env[63088]: _type = "HttpNfcLease" [ 1089.873377] env[63088]: } is initializing. {{(pid=63088) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1090.376533] env[63088]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1090.376533] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52fb97b9-8eca-b3e6-1723-ae511c9202ed" [ 1090.376533] env[63088]: _type = "HttpNfcLease" [ 1090.376533] env[63088]: } is ready. {{(pid=63088) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1090.376837] env[63088]: DEBUG oslo_vmware.rw_handles [None req-9e719bdd-5361-4621-a66f-918aaed14f4c tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1090.376837] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52fb97b9-8eca-b3e6-1723-ae511c9202ed" [ 1090.376837] env[63088]: _type = "HttpNfcLease" [ 1090.376837] env[63088]: }. {{(pid=63088) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1090.377586] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcd4011e-007f-408d-a69c-1a1c65f61ce7 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.384345] env[63088]: DEBUG oslo_vmware.rw_handles [None req-9e719bdd-5361-4621-a66f-918aaed14f4c tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/524237a1-5a6b-6d31-78c4-5cb9544d8bdd/disk-0.vmdk from lease info. {{(pid=63088) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1090.384685] env[63088]: DEBUG oslo_vmware.rw_handles [None req-9e719bdd-5361-4621-a66f-918aaed14f4c tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/524237a1-5a6b-6d31-78c4-5cb9544d8bdd/disk-0.vmdk for reading. {{(pid=63088) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1090.474046] env[63088]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-8333bda1-2072-462c-b7ca-76c4c2d7f408 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.605080] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1092.605080] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Cleaning up deleted instances {{(pid=63088) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11226}} [ 1093.118907] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] There are 48 instances to clean {{(pid=63088) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11235}} [ 1093.119209] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: 9d2e0a07-3542-4e05-9902-82a949baa5a8] Instance has had 0 of 5 cleanup attempts {{(pid=63088) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1093.623282] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: d94e57a3-743b-4bfe-b208-80f5748f8654] Instance has had 0 of 5 cleanup attempts {{(pid=63088) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1094.127382] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: b3b1a32f-9cc1-4810-9984-fd24c40c81cd] Instance has had 0 of 5 cleanup attempts {{(pid=63088) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1094.631152] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: 32418785-8823-4a5e-90a9-e205752530b7] Instance has had 0 of 5 cleanup attempts {{(pid=63088) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1095.134777] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: a39840c1-3d60-478b-987c-2519551963cf] Instance has had 0 of 5 cleanup attempts {{(pid=63088) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1095.639068] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: c4fb31b2-1301-4a6e-bef2-8ea462caee0d] Instance has had 0 of 5 cleanup attempts {{(pid=63088) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1096.142240] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: c41b1aa7-3cb8-4d33-8b98-612b96a8e0a3] Instance has had 0 of 5 cleanup attempts {{(pid=63088) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1096.646679] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: 13388466-4adb-4d56-9fc1-e3f5dc516077] Instance has had 0 of 5 cleanup attempts {{(pid=63088) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1097.149727] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: 69645f50-46fc-4c15-9b39-1721f7636e31] Instance has had 0 of 5 cleanup attempts {{(pid=63088) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1097.483786] env[63088]: DEBUG nova.compute.manager [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Stashing vm_state: active {{(pid=63088) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1097.643559] env[63088]: DEBUG oslo_vmware.rw_handles [None req-9e719bdd-5361-4621-a66f-918aaed14f4c tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/524237a1-5a6b-6d31-78c4-5cb9544d8bdd/disk-0.vmdk. {{(pid=63088) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1097.644515] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4ee396e-ec76-453f-9041-0386ab80c3c4 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.650396] env[63088]: DEBUG oslo_vmware.rw_handles [None req-9e719bdd-5361-4621-a66f-918aaed14f4c tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/524237a1-5a6b-6d31-78c4-5cb9544d8bdd/disk-0.vmdk is in state: ready. {{(pid=63088) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1097.650616] env[63088]: ERROR oslo_vmware.rw_handles [None req-9e719bdd-5361-4621-a66f-918aaed14f4c tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/524237a1-5a6b-6d31-78c4-5cb9544d8bdd/disk-0.vmdk due to incomplete transfer. [ 1097.650772] env[63088]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-032f73bf-e19e-4a22-b44b-87d7dc8538dd {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.652648] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: 66178bd0-7ddf-458c-98a2-3519e536dc63] Instance has had 0 of 5 cleanup attempts {{(pid=63088) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1097.659693] env[63088]: DEBUG oslo_vmware.rw_handles [None req-9e719bdd-5361-4621-a66f-918aaed14f4c tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/524237a1-5a6b-6d31-78c4-5cb9544d8bdd/disk-0.vmdk. {{(pid=63088) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1097.659888] env[63088]: DEBUG nova.virt.vmwareapi.images [None req-9e719bdd-5361-4621-a66f-918aaed14f4c tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Uploaded image ecd13d17-7e92-4cbe-805f-9c202f437a2f to the Glance image server {{(pid=63088) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1097.662349] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e719bdd-5361-4621-a66f-918aaed14f4c tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Destroying the VM {{(pid=63088) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1097.662562] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-4f0ad440-4451-4700-b01b-14e3d018cdcf {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.667552] env[63088]: DEBUG oslo_vmware.api [None req-9e719bdd-5361-4621-a66f-918aaed14f4c tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1097.667552] env[63088]: value = "task-1285532" [ 1097.667552] env[63088]: _type = "Task" [ 1097.667552] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.675106] env[63088]: DEBUG oslo_vmware.api [None req-9e719bdd-5361-4621-a66f-918aaed14f4c tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285532, 'name': Destroy_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.005081] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1098.005386] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1098.156260] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: 818704d4-2010-4492-8e51-28ad0bd6f570] Instance has had 0 of 5 cleanup attempts {{(pid=63088) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1098.177701] env[63088]: DEBUG oslo_vmware.api [None req-9e719bdd-5361-4621-a66f-918aaed14f4c tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285532, 'name': Destroy_Task, 'duration_secs': 0.296524} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.177954] env[63088]: INFO nova.virt.vmwareapi.vm_util [None req-9e719bdd-5361-4621-a66f-918aaed14f4c tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Destroyed the VM [ 1098.178216] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-9e719bdd-5361-4621-a66f-918aaed14f4c tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Deleting Snapshot of the VM instance {{(pid=63088) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1098.178469] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-0c3aa54c-1521-43b9-a6e5-f83988294968 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.185060] env[63088]: DEBUG oslo_vmware.api [None req-9e719bdd-5361-4621-a66f-918aaed14f4c tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1098.185060] env[63088]: value = "task-1285533" [ 1098.185060] env[63088]: _type = "Task" [ 1098.185060] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.192684] env[63088]: DEBUG oslo_vmware.api [None req-9e719bdd-5361-4621-a66f-918aaed14f4c tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285533, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.510711] env[63088]: INFO nova.compute.claims [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1098.659063] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: 4a56d74c-cb1f-4edf-99f0-4b54bf357a3a] Instance has had 0 of 5 cleanup attempts {{(pid=63088) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1098.695749] env[63088]: DEBUG oslo_vmware.api [None req-9e719bdd-5361-4621-a66f-918aaed14f4c tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285533, 'name': RemoveSnapshot_Task, 'duration_secs': 0.453851} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.696009] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-9e719bdd-5361-4621-a66f-918aaed14f4c tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Deleted Snapshot of the VM instance {{(pid=63088) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1098.696239] env[63088]: INFO nova.compute.manager [None req-9e719bdd-5361-4621-a66f-918aaed14f4c tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Took 11.96 seconds to snapshot the instance on the hypervisor. [ 1099.018057] env[63088]: INFO nova.compute.resource_tracker [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Updating resource usage from migration e77e7800-a0d0-4286-bebd-ba1e4a27bf70 [ 1099.071923] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d426cb21-ac0e-420f-8963-d5a69745c398 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.079452] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56d8cfc7-08e1-4794-b043-b30587baea92 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.107679] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a758b5d1-4592-43d1-bd5f-d439b4e72a5c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.114349] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6455fbb-2935-4988-b8e8-809866f0e7b0 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.132903] env[63088]: DEBUG nova.compute.provider_tree [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1099.162594] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: 49ee4180-6322-4003-a5b1-f2a91d190290] Instance has had 0 of 5 cleanup attempts {{(pid=63088) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1099.236913] env[63088]: DEBUG nova.compute.manager [None req-9e719bdd-5361-4621-a66f-918aaed14f4c tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Found 3 images (rotation: 2) {{(pid=63088) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 1099.237134] env[63088]: DEBUG nova.compute.manager [None req-9e719bdd-5361-4621-a66f-918aaed14f4c tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Rotating out 1 backups {{(pid=63088) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4562}} [ 1099.237306] env[63088]: DEBUG nova.compute.manager [None req-9e719bdd-5361-4621-a66f-918aaed14f4c tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Deleting image 7862f3ca-7713-4ff8-8498-b50e9658c73f {{(pid=63088) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4567}} [ 1099.636477] env[63088]: DEBUG nova.scheduler.client.report [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1099.665279] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: d6fef3ef-fec8-4929-b9b8-5e63306aa51d] Instance has had 0 of 5 cleanup attempts {{(pid=63088) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1100.141499] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.136s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1100.141675] env[63088]: INFO nova.compute.manager [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Migrating [ 1100.167961] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: 8205f9e1-effd-442d-990b-bf5a2a9e6cdd] Instance has had 0 of 5 cleanup attempts {{(pid=63088) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1100.658037] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquiring lock "refresh_cache-b25b5883-20c0-4cc9-8cc1-2c22a0f5be81" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1100.658037] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquired lock "refresh_cache-b25b5883-20c0-4cc9-8cc1-2c22a0f5be81" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1100.658037] env[63088]: DEBUG nova.network.neutron [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1100.670996] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: 7cd45e52-c6be-4ec6-9710-fa99e29cf872] Instance has had 0 of 5 cleanup attempts {{(pid=63088) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1101.174685] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: 2ad33254-8030-4454-b023-3359071f85de] Instance has had 0 of 5 cleanup attempts {{(pid=63088) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1101.342821] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a6323fc1-697c-4cd2-b5ee-64170072a336 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquiring lock "2b87a9f1-5f10-43c2-8bc9-6d560ec88015" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1101.343204] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a6323fc1-697c-4cd2-b5ee-64170072a336 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lock "2b87a9f1-5f10-43c2-8bc9-6d560ec88015" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1101.343332] env[63088]: DEBUG nova.compute.manager [None req-a6323fc1-697c-4cd2-b5ee-64170072a336 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1101.344247] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc7b3e22-e141-4843-b1e2-8e1bfaf7c073 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.351265] env[63088]: DEBUG nova.compute.manager [None req-a6323fc1-697c-4cd2-b5ee-64170072a336 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=63088) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1101.351799] env[63088]: DEBUG nova.objects.instance [None req-a6323fc1-697c-4cd2-b5ee-64170072a336 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lazy-loading 'flavor' on Instance uuid 2b87a9f1-5f10-43c2-8bc9-6d560ec88015 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1101.391619] env[63088]: DEBUG nova.network.neutron [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Updating instance_info_cache with network_info: [{"id": "f344e65c-4141-4c1d-a8e0-4aa1571a8e2f", "address": "fa:16:3e:3f:78:a4", "network": {"id": "dfcbe153-bc01-4362-9247-de9f0b1c847b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-841757397-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.152", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dace8b5181b84623b08f903d12dfd31e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c68b7663-4f0e-47f0-ac7f-40c6d952f7bb", "external-id": "nsx-vlan-transportzone-696", "segmentation_id": 696, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf344e65c-41", "ovs_interfaceid": "f344e65c-4141-4c1d-a8e0-4aa1571a8e2f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1101.679602] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: 79c4d6e8-9999-4b0c-98d3-bbfd4d61e660] Instance has had 0 of 5 cleanup attempts {{(pid=63088) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1101.858240] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6323fc1-697c-4cd2-b5ee-64170072a336 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1101.858490] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6825334f-5377-45ac-b480-1da21ae386f1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.866325] env[63088]: DEBUG oslo_vmware.api [None req-a6323fc1-697c-4cd2-b5ee-64170072a336 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1101.866325] env[63088]: value = "task-1285534" [ 1101.866325] env[63088]: _type = "Task" [ 1101.866325] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.874235] env[63088]: DEBUG oslo_vmware.api [None req-a6323fc1-697c-4cd2-b5ee-64170072a336 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285534, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.895195] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Releasing lock "refresh_cache-b25b5883-20c0-4cc9-8cc1-2c22a0f5be81" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1102.183180] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: f582da15-dfc3-45e1-a995-9dd0c9533869] Instance has had 0 of 5 cleanup attempts {{(pid=63088) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1102.375815] env[63088]: DEBUG oslo_vmware.api [None req-a6323fc1-697c-4cd2-b5ee-64170072a336 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285534, 'name': PowerOffVM_Task, 'duration_secs': 0.178833} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.376080] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6323fc1-697c-4cd2-b5ee-64170072a336 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1102.376263] env[63088]: DEBUG nova.compute.manager [None req-a6323fc1-697c-4cd2-b5ee-64170072a336 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1102.377026] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f2c01bb-d91d-4f08-859b-000bf49730d4 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.686591] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: fa9d1ee9-18c6-44d8-b80f-d7390f3e49d3] Instance has had 0 of 5 cleanup attempts {{(pid=63088) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1102.887601] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a6323fc1-697c-4cd2-b5ee-64170072a336 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lock "2b87a9f1-5f10-43c2-8bc9-6d560ec88015" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.544s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1103.190627] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: 45513c6b-c000-4ee1-8893-4e084ffc22c5] Instance has had 0 of 5 cleanup attempts {{(pid=63088) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1103.410956] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64078ac8-9493-4240-a0b7-c201fd3869af {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.427511] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Updating instance 'b25b5883-20c0-4cc9-8cc1-2c22a0f5be81' progress to 0 {{(pid=63088) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1103.694168] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: bebc3318-24bb-4a37-8b23-66a12a7f7fd2] Instance has had 0 of 5 cleanup attempts {{(pid=63088) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1103.934609] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1103.934885] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ca55d5f2-57cb-4ecd-ad2b-559b491da9f4 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.941592] env[63088]: DEBUG oslo_vmware.api [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for the task: (returnval){ [ 1103.941592] env[63088]: value = "task-1285535" [ 1103.941592] env[63088]: _type = "Task" [ 1103.941592] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.949254] env[63088]: DEBUG oslo_vmware.api [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285535, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.196855] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: 71e37d8e-a454-46c4-a3cc-3d5671a32beb] Instance has had 0 of 5 cleanup attempts {{(pid=63088) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1104.452894] env[63088]: DEBUG oslo_vmware.api [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285535, 'name': PowerOffVM_Task, 'duration_secs': 0.164298} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.453152] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1104.453356] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Updating instance 'b25b5883-20c0-4cc9-8cc1-2c22a0f5be81' progress to 17 {{(pid=63088) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1104.627449] env[63088]: DEBUG nova.compute.manager [None req-af441786-2d51-4cbf-ab87-ac35d2d4a26e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Stashing vm_state: stopped {{(pid=63088) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1104.700227] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: 254db932-35f1-42e3-9207-cd886efd65a3] Instance has had 0 of 5 cleanup attempts {{(pid=63088) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1104.959711] env[63088]: DEBUG nova.virt.hardware [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:33Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1104.959938] env[63088]: DEBUG nova.virt.hardware [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1104.960166] env[63088]: DEBUG nova.virt.hardware [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1104.960299] env[63088]: DEBUG nova.virt.hardware [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1104.960476] env[63088]: DEBUG nova.virt.hardware [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1104.960600] env[63088]: DEBUG nova.virt.hardware [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1104.960833] env[63088]: DEBUG nova.virt.hardware [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1104.960979] env[63088]: DEBUG nova.virt.hardware [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1104.961152] env[63088]: DEBUG nova.virt.hardware [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1104.961314] env[63088]: DEBUG nova.virt.hardware [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1104.961492] env[63088]: DEBUG nova.virt.hardware [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1104.967134] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1980ed48-e298-4d07-9f7e-55f752959faf {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.982714] env[63088]: DEBUG oslo_vmware.api [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for the task: (returnval){ [ 1104.982714] env[63088]: value = "task-1285536" [ 1104.982714] env[63088]: _type = "Task" [ 1104.982714] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.990396] env[63088]: DEBUG oslo_vmware.api [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285536, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.203857] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: 2964db35-7357-40a7-b4e6-7e2595549f5b] Instance has had 0 of 5 cleanup attempts {{(pid=63088) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1105.219364] env[63088]: DEBUG oslo_concurrency.lockutils [None req-af441786-2d51-4cbf-ab87-ac35d2d4a26e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1105.219460] env[63088]: DEBUG oslo_concurrency.lockutils [None req-af441786-2d51-4cbf-ab87-ac35d2d4a26e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1105.493101] env[63088]: DEBUG oslo_vmware.api [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285536, 'name': ReconfigVM_Task, 'duration_secs': 0.164528} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.493368] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Updating instance 'b25b5883-20c0-4cc9-8cc1-2c22a0f5be81' progress to 33 {{(pid=63088) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1105.707307] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: b50db882-598c-488f-b935-34f55f655642] Instance has had 0 of 5 cleanup attempts {{(pid=63088) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1105.725166] env[63088]: INFO nova.compute.claims [None req-af441786-2d51-4cbf-ab87-ac35d2d4a26e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1106.000146] env[63088]: DEBUG nova.virt.hardware [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1106.000424] env[63088]: DEBUG nova.virt.hardware [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1106.000624] env[63088]: DEBUG nova.virt.hardware [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1106.000822] env[63088]: DEBUG nova.virt.hardware [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1106.000973] env[63088]: DEBUG nova.virt.hardware [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1106.001142] env[63088]: DEBUG nova.virt.hardware [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1106.001353] env[63088]: DEBUG nova.virt.hardware [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1106.001525] env[63088]: DEBUG nova.virt.hardware [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1106.001730] env[63088]: DEBUG nova.virt.hardware [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1106.001902] env[63088]: DEBUG nova.virt.hardware [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1106.002105] env[63088]: DEBUG nova.virt.hardware [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1106.007459] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Reconfiguring VM instance instance-00000067 to detach disk 2000 {{(pid=63088) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1106.008118] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-49cdd2e4-0717-4e8f-8cd5-35bd358b8600 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.027573] env[63088]: DEBUG oslo_vmware.api [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for the task: (returnval){ [ 1106.027573] env[63088]: value = "task-1285537" [ 1106.027573] env[63088]: _type = "Task" [ 1106.027573] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.035433] env[63088]: DEBUG oslo_vmware.api [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285537, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.211194] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: 667fd5e9-5fe4-41e1-9d8a-896c6e5c6286] Instance has had 0 of 5 cleanup attempts {{(pid=63088) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1106.231070] env[63088]: INFO nova.compute.resource_tracker [None req-af441786-2d51-4cbf-ab87-ac35d2d4a26e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Updating resource usage from migration 74ec533f-3231-45f3-93f2-fa9c6481ffc0 [ 1106.293915] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fb66ab0-4b46-400a-bf49-b172ac6a0575 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.301508] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03ba5a9c-a4a7-4b60-98df-c981e769665a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.331953] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec0bbdc9-9274-45be-99c3-82f696203ec1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.338859] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96af0ec8-c53b-405f-8b96-6188d262128a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.351359] env[63088]: DEBUG nova.compute.provider_tree [None req-af441786-2d51-4cbf-ab87-ac35d2d4a26e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1106.536771] env[63088]: DEBUG oslo_vmware.api [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285537, 'name': ReconfigVM_Task, 'duration_secs': 0.165574} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.537062] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Reconfigured VM instance instance-00000067 to detach disk 2000 {{(pid=63088) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1106.537815] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8633ae53-7d7e-48bb-b2b1-a546d45bb13c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.558838] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Reconfiguring VM instance instance-00000067 to attach disk [datastore2] b25b5883-20c0-4cc9-8cc1-2c22a0f5be81/b25b5883-20c0-4cc9-8cc1-2c22a0f5be81.vmdk or device None with type thin {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1106.559090] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-53e30788-bac0-4788-bcfc-3071c2d76ea0 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.575830] env[63088]: DEBUG oslo_vmware.api [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for the task: (returnval){ [ 1106.575830] env[63088]: value = "task-1285538" [ 1106.575830] env[63088]: _type = "Task" [ 1106.575830] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.583072] env[63088]: DEBUG oslo_vmware.api [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285538, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.714491] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: 0fdb21d1-4111-4ff3-bdc0-e2598298a9a4] Instance has had 0 of 5 cleanup attempts {{(pid=63088) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1106.855061] env[63088]: DEBUG nova.scheduler.client.report [None req-af441786-2d51-4cbf-ab87-ac35d2d4a26e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1107.085251] env[63088]: DEBUG oslo_vmware.api [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285538, 'name': ReconfigVM_Task, 'duration_secs': 0.266979} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.085528] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Reconfigured VM instance instance-00000067 to attach disk [datastore2] b25b5883-20c0-4cc9-8cc1-2c22a0f5be81/b25b5883-20c0-4cc9-8cc1-2c22a0f5be81.vmdk or device None with type thin {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1107.085771] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Updating instance 'b25b5883-20c0-4cc9-8cc1-2c22a0f5be81' progress to 50 {{(pid=63088) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1107.218539] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: 2269e520-08b6-433f-8fe9-7b84d91e02d7] Instance has had 0 of 5 cleanup attempts {{(pid=63088) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1107.359576] env[63088]: DEBUG oslo_concurrency.lockutils [None req-af441786-2d51-4cbf-ab87-ac35d2d4a26e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.140s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1107.359789] env[63088]: INFO nova.compute.manager [None req-af441786-2d51-4cbf-ab87-ac35d2d4a26e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Migrating [ 1107.592258] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ca46613-bed3-4974-8b74-23bfb63e181e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.611689] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c30ce44-dc66-4cbe-a34e-52e5bcbc2150 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.628288] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Updating instance 'b25b5883-20c0-4cc9-8cc1-2c22a0f5be81' progress to 67 {{(pid=63088) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1107.721883] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: 9d5f1cde-e787-4a2c-ac49-83ec135ff51c] Instance has had 0 of 5 cleanup attempts {{(pid=63088) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1107.874296] env[63088]: DEBUG oslo_concurrency.lockutils [None req-af441786-2d51-4cbf-ab87-ac35d2d4a26e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquiring lock "refresh_cache-2b87a9f1-5f10-43c2-8bc9-6d560ec88015" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1107.874442] env[63088]: DEBUG oslo_concurrency.lockutils [None req-af441786-2d51-4cbf-ab87-ac35d2d4a26e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquired lock "refresh_cache-2b87a9f1-5f10-43c2-8bc9-6d560ec88015" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1107.874628] env[63088]: DEBUG nova.network.neutron [None req-af441786-2d51-4cbf-ab87-ac35d2d4a26e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1108.168152] env[63088]: DEBUG nova.network.neutron [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Port f344e65c-4141-4c1d-a8e0-4aa1571a8e2f binding to destination host cpu-1 is already ACTIVE {{(pid=63088) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1108.225519] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: 58d561fe-26aa-4e94-8d55-cc70c361b479] Instance has had 0 of 5 cleanup attempts {{(pid=63088) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1108.599373] env[63088]: DEBUG nova.network.neutron [None req-af441786-2d51-4cbf-ab87-ac35d2d4a26e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Updating instance_info_cache with network_info: [{"id": "f88388b7-bc70-4095-a792-8d155ef34534", "address": "fa:16:3e:f7:d1:c7", "network": {"id": "711d3d9a-86db-4aa1-87b5-481f1812dfa4", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-59376186-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.225", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a7898ddafe0d41038a6ae4277f2c2f48", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e41070eb-3ac1-4ca9-a3d0-fd65893a97de", "external-id": "nsx-vlan-transportzone-596", "segmentation_id": 596, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf88388b7-bc", "ovs_interfaceid": "f88388b7-bc70-4095-a792-8d155ef34534", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1108.728987] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: e6b0ce08-d67f-458f-92f5-1e904d03a5ef] Instance has had 0 of 5 cleanup attempts {{(pid=63088) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1109.102646] env[63088]: DEBUG oslo_concurrency.lockutils [None req-af441786-2d51-4cbf-ab87-ac35d2d4a26e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Releasing lock "refresh_cache-2b87a9f1-5f10-43c2-8bc9-6d560ec88015" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1109.189407] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquiring lock "b25b5883-20c0-4cc9-8cc1-2c22a0f5be81-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1109.189715] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lock "b25b5883-20c0-4cc9-8cc1-2c22a0f5be81-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1109.189944] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lock "b25b5883-20c0-4cc9-8cc1-2c22a0f5be81-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1109.232142] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: dc4b2c2c-a3f3-4786-9f34-45f8bcacfb69] Instance has had 0 of 5 cleanup attempts {{(pid=63088) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1109.735442] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: 7b6aadb7-e34b-42b7-b69f-370434f5b665] Instance has had 0 of 5 cleanup attempts {{(pid=63088) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1110.224814] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquiring lock "refresh_cache-b25b5883-20c0-4cc9-8cc1-2c22a0f5be81" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1110.225074] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquired lock "refresh_cache-b25b5883-20c0-4cc9-8cc1-2c22a0f5be81" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1110.225265] env[63088]: DEBUG nova.network.neutron [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1110.238057] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: 8b564e29-96cf-4abf-963d-142b413fb464] Instance has had 0 of 5 cleanup attempts {{(pid=63088) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1110.617647] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddd6deb8-c422-4551-b838-9a6ec0dd3332 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.636789] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-af441786-2d51-4cbf-ab87-ac35d2d4a26e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Updating instance '2b87a9f1-5f10-43c2-8bc9-6d560ec88015' progress to 0 {{(pid=63088) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1110.740342] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: 7f341d80-450a-47b8-b26d-15f2ce3e378c] Instance has had 0 of 5 cleanup attempts {{(pid=63088) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1110.799601] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Acquiring lock "3fa11db4-2f70-495e-a562-ac5f06153847" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1110.799959] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Lock "3fa11db4-2f70-495e-a562-ac5f06153847" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1110.800240] env[63088]: INFO nova.compute.manager [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Shelving [ 1110.925123] env[63088]: DEBUG nova.network.neutron [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Updating instance_info_cache with network_info: [{"id": "f344e65c-4141-4c1d-a8e0-4aa1571a8e2f", "address": "fa:16:3e:3f:78:a4", "network": {"id": "dfcbe153-bc01-4362-9247-de9f0b1c847b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-841757397-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.152", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dace8b5181b84623b08f903d12dfd31e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c68b7663-4f0e-47f0-ac7f-40c6d952f7bb", "external-id": "nsx-vlan-transportzone-696", "segmentation_id": 696, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf344e65c-41", "ovs_interfaceid": "f344e65c-4141-4c1d-a8e0-4aa1571a8e2f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1111.142292] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-af441786-2d51-4cbf-ab87-ac35d2d4a26e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1111.142562] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e02ba27d-9093-4b8a-b995-09fc03d6eecb {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.150547] env[63088]: DEBUG oslo_vmware.api [None req-af441786-2d51-4cbf-ab87-ac35d2d4a26e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1111.150547] env[63088]: value = "task-1285539" [ 1111.150547] env[63088]: _type = "Task" [ 1111.150547] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.158508] env[63088]: DEBUG oslo_vmware.api [None req-af441786-2d51-4cbf-ab87-ac35d2d4a26e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285539, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.243213] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: 00925f94-dbf9-453c-a124-d8434679aedf] Instance has had 0 of 5 cleanup attempts {{(pid=63088) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1111.307018] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1111.307333] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5d533b39-5c0c-40bf-bfe2-d26f8337a742 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.314469] env[63088]: DEBUG oslo_vmware.api [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Waiting for the task: (returnval){ [ 1111.314469] env[63088]: value = "task-1285540" [ 1111.314469] env[63088]: _type = "Task" [ 1111.314469] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.323132] env[63088]: DEBUG oslo_vmware.api [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285540, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.427642] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Releasing lock "refresh_cache-b25b5883-20c0-4cc9-8cc1-2c22a0f5be81" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1111.660848] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-af441786-2d51-4cbf-ab87-ac35d2d4a26e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] VM already powered off {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1111.661098] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-af441786-2d51-4cbf-ab87-ac35d2d4a26e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Updating instance '2b87a9f1-5f10-43c2-8bc9-6d560ec88015' progress to 17 {{(pid=63088) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1111.746369] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: 2c2b8d1f-bb3a-49e1-8d28-00eae4f00a68] Instance has had 0 of 5 cleanup attempts {{(pid=63088) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1111.823447] env[63088]: DEBUG oslo_vmware.api [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285540, 'name': PowerOffVM_Task, 'duration_secs': 0.163529} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.823726] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1111.824481] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7c72a3f-07ef-469d-b4f7-3085bbfeeaf3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.842993] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3c9c275-d205-4aca-bacb-613aee5f4ccd {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.952400] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c24b822-3389-410b-b1b3-8bcc0d284244 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.971041] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8230bc13-d45f-4d3f-878e-ed2e570cf775 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.977718] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Updating instance 'b25b5883-20c0-4cc9-8cc1-2c22a0f5be81' progress to 83 {{(pid=63088) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1112.166947] env[63088]: DEBUG nova.virt.hardware [None req-af441786-2d51-4cbf-ab87-ac35d2d4a26e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:33Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1112.167204] env[63088]: DEBUG nova.virt.hardware [None req-af441786-2d51-4cbf-ab87-ac35d2d4a26e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1112.167381] env[63088]: DEBUG nova.virt.hardware [None req-af441786-2d51-4cbf-ab87-ac35d2d4a26e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1112.167572] env[63088]: DEBUG nova.virt.hardware [None req-af441786-2d51-4cbf-ab87-ac35d2d4a26e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1112.167725] env[63088]: DEBUG nova.virt.hardware [None req-af441786-2d51-4cbf-ab87-ac35d2d4a26e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1112.167876] env[63088]: DEBUG nova.virt.hardware [None req-af441786-2d51-4cbf-ab87-ac35d2d4a26e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1112.168121] env[63088]: DEBUG nova.virt.hardware [None req-af441786-2d51-4cbf-ab87-ac35d2d4a26e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1112.168291] env[63088]: DEBUG nova.virt.hardware [None req-af441786-2d51-4cbf-ab87-ac35d2d4a26e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1112.168474] env[63088]: DEBUG nova.virt.hardware [None req-af441786-2d51-4cbf-ab87-ac35d2d4a26e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1112.168649] env[63088]: DEBUG nova.virt.hardware [None req-af441786-2d51-4cbf-ab87-ac35d2d4a26e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1112.168828] env[63088]: DEBUG nova.virt.hardware [None req-af441786-2d51-4cbf-ab87-ac35d2d4a26e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1112.173787] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e39fe747-4793-4dd5-a1ad-16bcdfa00ef0 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.188646] env[63088]: DEBUG oslo_vmware.api [None req-af441786-2d51-4cbf-ab87-ac35d2d4a26e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1112.188646] env[63088]: value = "task-1285541" [ 1112.188646] env[63088]: _type = "Task" [ 1112.188646] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.196838] env[63088]: DEBUG oslo_vmware.api [None req-af441786-2d51-4cbf-ab87-ac35d2d4a26e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285541, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.249534] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: 1d421bf0-925d-408c-a612-b659da92799f] Instance has had 0 of 5 cleanup attempts {{(pid=63088) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1112.352291] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Creating Snapshot of the VM instance {{(pid=63088) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1112.353033] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-0fded012-1afa-46fc-8650-8f72c617c84d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.360565] env[63088]: DEBUG oslo_vmware.api [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Waiting for the task: (returnval){ [ 1112.360565] env[63088]: value = "task-1285542" [ 1112.360565] env[63088]: _type = "Task" [ 1112.360565] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.369207] env[63088]: DEBUG oslo_vmware.api [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285542, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.484328] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1112.484588] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-38ab29d2-92f3-40fd-bcec-115c13d55a71 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.491325] env[63088]: DEBUG oslo_vmware.api [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for the task: (returnval){ [ 1112.491325] env[63088]: value = "task-1285543" [ 1112.491325] env[63088]: _type = "Task" [ 1112.491325] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.499237] env[63088]: DEBUG oslo_vmware.api [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285543, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.697886] env[63088]: DEBUG oslo_vmware.api [None req-af441786-2d51-4cbf-ab87-ac35d2d4a26e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285541, 'name': ReconfigVM_Task, 'duration_secs': 0.136136} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.698320] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-af441786-2d51-4cbf-ab87-ac35d2d4a26e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Updating instance '2b87a9f1-5f10-43c2-8bc9-6d560ec88015' progress to 33 {{(pid=63088) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1112.753086] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: 34ce411f-40c8-446e-b685-cd3ce07663dd] Instance has had 0 of 5 cleanup attempts {{(pid=63088) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1112.869467] env[63088]: DEBUG oslo_vmware.api [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285542, 'name': CreateSnapshot_Task, 'duration_secs': 0.438505} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.869724] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Created Snapshot of the VM instance {{(pid=63088) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1112.870465] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5b229e7-d841-4815-ad68-b3168b8a90fb {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.002766] env[63088]: DEBUG oslo_vmware.api [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285543, 'name': PowerOnVM_Task, 'duration_secs': 0.370742} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.003074] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1113.003276] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-5aad946d-a46a-4c40-bdf8-2d71a5a5fd6f tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Updating instance 'b25b5883-20c0-4cc9-8cc1-2c22a0f5be81' progress to 100 {{(pid=63088) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1113.204860] env[63088]: DEBUG nova.virt.hardware [None req-af441786-2d51-4cbf-ab87-ac35d2d4a26e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1113.205174] env[63088]: DEBUG nova.virt.hardware [None req-af441786-2d51-4cbf-ab87-ac35d2d4a26e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1113.205341] env[63088]: DEBUG nova.virt.hardware [None req-af441786-2d51-4cbf-ab87-ac35d2d4a26e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1113.205547] env[63088]: DEBUG nova.virt.hardware [None req-af441786-2d51-4cbf-ab87-ac35d2d4a26e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1113.205684] env[63088]: DEBUG nova.virt.hardware [None req-af441786-2d51-4cbf-ab87-ac35d2d4a26e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1113.205837] env[63088]: DEBUG nova.virt.hardware [None req-af441786-2d51-4cbf-ab87-ac35d2d4a26e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1113.206098] env[63088]: DEBUG nova.virt.hardware [None req-af441786-2d51-4cbf-ab87-ac35d2d4a26e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1113.206284] env[63088]: DEBUG nova.virt.hardware [None req-af441786-2d51-4cbf-ab87-ac35d2d4a26e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1113.206460] env[63088]: DEBUG nova.virt.hardware [None req-af441786-2d51-4cbf-ab87-ac35d2d4a26e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1113.207020] env[63088]: DEBUG nova.virt.hardware [None req-af441786-2d51-4cbf-ab87-ac35d2d4a26e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1113.207020] env[63088]: DEBUG nova.virt.hardware [None req-af441786-2d51-4cbf-ab87-ac35d2d4a26e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1113.212236] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-af441786-2d51-4cbf-ab87-ac35d2d4a26e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Reconfiguring VM instance instance-00000063 to detach disk 2000 {{(pid=63088) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1113.212521] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f93b6e23-73f2-4159-ba0a-399d82ba3751 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.230792] env[63088]: DEBUG oslo_vmware.api [None req-af441786-2d51-4cbf-ab87-ac35d2d4a26e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1113.230792] env[63088]: value = "task-1285544" [ 1113.230792] env[63088]: _type = "Task" [ 1113.230792] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.238495] env[63088]: DEBUG oslo_vmware.api [None req-af441786-2d51-4cbf-ab87-ac35d2d4a26e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285544, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.257111] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: b159649b-ebd5-4c7d-9074-dc0e6395a947] Instance has had 0 of 5 cleanup attempts {{(pid=63088) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1113.387296] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Creating linked-clone VM from snapshot {{(pid=63088) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1113.387621] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-e1a5043b-c94b-4fcb-940e-4806dfa51cb5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.396342] env[63088]: DEBUG oslo_vmware.api [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Waiting for the task: (returnval){ [ 1113.396342] env[63088]: value = "task-1285545" [ 1113.396342] env[63088]: _type = "Task" [ 1113.396342] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.403569] env[63088]: DEBUG oslo_vmware.api [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285545, 'name': CloneVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.740202] env[63088]: DEBUG oslo_vmware.api [None req-af441786-2d51-4cbf-ab87-ac35d2d4a26e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285544, 'name': ReconfigVM_Task, 'duration_secs': 0.148961} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.740491] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-af441786-2d51-4cbf-ab87-ac35d2d4a26e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Reconfigured VM instance instance-00000063 to detach disk 2000 {{(pid=63088) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1113.741275] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3f2a3d7-c9fd-40d3-9bd8-30da83594777 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.762350] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-af441786-2d51-4cbf-ab87-ac35d2d4a26e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Reconfiguring VM instance instance-00000063 to attach disk [datastore1] 2b87a9f1-5f10-43c2-8bc9-6d560ec88015/2b87a9f1-5f10-43c2-8bc9-6d560ec88015.vmdk or device None with type thin {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1113.762787] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: 07fd3e6f-e0af-467c-9039-238be5c58d25] Instance has had 0 of 5 cleanup attempts {{(pid=63088) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1113.764456] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ab811059-d475-4f00-bdad-4289f68fe2c6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.783339] env[63088]: DEBUG oslo_vmware.api [None req-af441786-2d51-4cbf-ab87-ac35d2d4a26e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1113.783339] env[63088]: value = "task-1285546" [ 1113.783339] env[63088]: _type = "Task" [ 1113.783339] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.791390] env[63088]: DEBUG oslo_vmware.api [None req-af441786-2d51-4cbf-ab87-ac35d2d4a26e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285546, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.906868] env[63088]: DEBUG oslo_vmware.api [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285545, 'name': CloneVM_Task} progress is 94%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.278290] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: 4f6e2d99-7300-42e6-8abf-362ddd4e8c6d] Instance has had 0 of 5 cleanup attempts {{(pid=63088) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1114.292622] env[63088]: DEBUG oslo_vmware.api [None req-af441786-2d51-4cbf-ab87-ac35d2d4a26e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285546, 'name': ReconfigVM_Task, 'duration_secs': 0.243108} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.293079] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-af441786-2d51-4cbf-ab87-ac35d2d4a26e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Reconfigured VM instance instance-00000063 to attach disk [datastore1] 2b87a9f1-5f10-43c2-8bc9-6d560ec88015/2b87a9f1-5f10-43c2-8bc9-6d560ec88015.vmdk or device None with type thin {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1114.293225] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-af441786-2d51-4cbf-ab87-ac35d2d4a26e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Updating instance '2b87a9f1-5f10-43c2-8bc9-6d560ec88015' progress to 50 {{(pid=63088) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1114.408358] env[63088]: DEBUG oslo_vmware.api [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285545, 'name': CloneVM_Task} progress is 95%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.782008] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: e845f60b-004f-4a5c-813e-0c183e99eba6] Instance has had 0 of 5 cleanup attempts {{(pid=63088) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1114.801868] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67a0a150-573c-4261-a53a-39096e6341e0 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.824594] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5498c5c-74be-46a0-ad59-c15b640c1dbd {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.844429] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-af441786-2d51-4cbf-ab87-ac35d2d4a26e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Updating instance '2b87a9f1-5f10-43c2-8bc9-6d560ec88015' progress to 67 {{(pid=63088) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1114.896104] env[63088]: DEBUG nova.network.neutron [None req-60ee2949-6f9b-4a40-8984-0f26595d38e8 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Port f344e65c-4141-4c1d-a8e0-4aa1571a8e2f binding to destination host cpu-1 is already ACTIVE {{(pid=63088) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1114.896417] env[63088]: DEBUG oslo_concurrency.lockutils [None req-60ee2949-6f9b-4a40-8984-0f26595d38e8 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquiring lock "refresh_cache-b25b5883-20c0-4cc9-8cc1-2c22a0f5be81" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1114.896596] env[63088]: DEBUG oslo_concurrency.lockutils [None req-60ee2949-6f9b-4a40-8984-0f26595d38e8 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquired lock "refresh_cache-b25b5883-20c0-4cc9-8cc1-2c22a0f5be81" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1114.896779] env[63088]: DEBUG nova.network.neutron [None req-60ee2949-6f9b-4a40-8984-0f26595d38e8 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1114.907788] env[63088]: DEBUG oslo_vmware.api [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285545, 'name': CloneVM_Task, 'duration_secs': 1.207141} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.908072] env[63088]: INFO nova.virt.vmwareapi.vmops [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Created linked-clone VM from snapshot [ 1114.908817] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c34ca748-32a7-4bd7-aba5-125dbf5b51fe {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.916432] env[63088]: DEBUG nova.virt.vmwareapi.images [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Uploading image f1db6e74-0b10-490d-a958-6fa032fe3caf {{(pid=63088) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1114.943095] env[63088]: DEBUG oslo_vmware.rw_handles [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1114.943095] env[63088]: value = "vm-275965" [ 1114.943095] env[63088]: _type = "VirtualMachine" [ 1114.943095] env[63088]: }. {{(pid=63088) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1114.943713] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-f1b1ad8f-f02d-4ce4-975e-9b009a550051 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.950400] env[63088]: DEBUG oslo_vmware.rw_handles [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Lease: (returnval){ [ 1114.950400] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]525f227e-81b0-b9a6-594d-86d1e22009f4" [ 1114.950400] env[63088]: _type = "HttpNfcLease" [ 1114.950400] env[63088]: } obtained for exporting VM: (result){ [ 1114.950400] env[63088]: value = "vm-275965" [ 1114.950400] env[63088]: _type = "VirtualMachine" [ 1114.950400] env[63088]: }. {{(pid=63088) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1114.950651] env[63088]: DEBUG oslo_vmware.api [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Waiting for the lease: (returnval){ [ 1114.950651] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]525f227e-81b0-b9a6-594d-86d1e22009f4" [ 1114.950651] env[63088]: _type = "HttpNfcLease" [ 1114.950651] env[63088]: } to be ready. {{(pid=63088) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1114.957364] env[63088]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1114.957364] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]525f227e-81b0-b9a6-594d-86d1e22009f4" [ 1114.957364] env[63088]: _type = "HttpNfcLease" [ 1114.957364] env[63088]: } is initializing. {{(pid=63088) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1115.285918] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: 4847f15e-98d4-401c-91ff-067e84a85727] Instance has had 0 of 5 cleanup attempts {{(pid=63088) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1115.391839] env[63088]: DEBUG nova.network.neutron [None req-af441786-2d51-4cbf-ab87-ac35d2d4a26e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Port f88388b7-bc70-4095-a792-8d155ef34534 binding to destination host cpu-1 is already ACTIVE {{(pid=63088) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1115.461255] env[63088]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1115.461255] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]525f227e-81b0-b9a6-594d-86d1e22009f4" [ 1115.461255] env[63088]: _type = "HttpNfcLease" [ 1115.461255] env[63088]: } is ready. {{(pid=63088) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1115.463261] env[63088]: DEBUG oslo_vmware.rw_handles [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1115.463261] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]525f227e-81b0-b9a6-594d-86d1e22009f4" [ 1115.463261] env[63088]: _type = "HttpNfcLease" [ 1115.463261] env[63088]: }. {{(pid=63088) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1115.464026] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b103da6-94c5-460d-94ec-97b95963f78c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.472782] env[63088]: DEBUG oslo_vmware.rw_handles [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c35194-26f1-f149-0043-900414536f3c/disk-0.vmdk from lease info. {{(pid=63088) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1115.472968] env[63088]: DEBUG oslo_vmware.rw_handles [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c35194-26f1-f149-0043-900414536f3c/disk-0.vmdk for reading. {{(pid=63088) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1115.572885] env[63088]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-cddc8cb7-889e-4642-9062-52b6b203edc8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.643940] env[63088]: DEBUG nova.network.neutron [None req-60ee2949-6f9b-4a40-8984-0f26595d38e8 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Updating instance_info_cache with network_info: [{"id": "f344e65c-4141-4c1d-a8e0-4aa1571a8e2f", "address": "fa:16:3e:3f:78:a4", "network": {"id": "dfcbe153-bc01-4362-9247-de9f0b1c847b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-841757397-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.152", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dace8b5181b84623b08f903d12dfd31e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c68b7663-4f0e-47f0-ac7f-40c6d952f7bb", "external-id": "nsx-vlan-transportzone-696", "segmentation_id": 696, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf344e65c-41", "ovs_interfaceid": "f344e65c-4141-4c1d-a8e0-4aa1571a8e2f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1115.789090] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: 4971b24c-6710-4f50-9846-727dad264b1f] Instance has had 0 of 5 cleanup attempts {{(pid=63088) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1116.146987] env[63088]: DEBUG oslo_concurrency.lockutils [None req-60ee2949-6f9b-4a40-8984-0f26595d38e8 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Releasing lock "refresh_cache-b25b5883-20c0-4cc9-8cc1-2c22a0f5be81" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1116.293286] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: 11ec987f-9908-4366-81ae-cbfe4de67b2d] Instance has had 0 of 5 cleanup attempts {{(pid=63088) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1116.408989] env[63088]: DEBUG oslo_concurrency.lockutils [None req-af441786-2d51-4cbf-ab87-ac35d2d4a26e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquiring lock "2b87a9f1-5f10-43c2-8bc9-6d560ec88015-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1116.410024] env[63088]: DEBUG oslo_concurrency.lockutils [None req-af441786-2d51-4cbf-ab87-ac35d2d4a26e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lock "2b87a9f1-5f10-43c2-8bc9-6d560ec88015-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1116.410024] env[63088]: DEBUG oslo_concurrency.lockutils [None req-af441786-2d51-4cbf-ab87-ac35d2d4a26e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lock "2b87a9f1-5f10-43c2-8bc9-6d560ec88015-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1116.650720] env[63088]: DEBUG nova.compute.manager [None req-60ee2949-6f9b-4a40-8984-0f26595d38e8 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=63088) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:897}} [ 1116.651075] env[63088]: DEBUG oslo_concurrency.lockutils [None req-60ee2949-6f9b-4a40-8984-0f26595d38e8 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1116.651393] env[63088]: DEBUG oslo_concurrency.lockutils [None req-60ee2949-6f9b-4a40-8984-0f26595d38e8 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1116.797859] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: db032da3-d334-486f-a428-af1c8a3c360a] Instance has had 0 of 5 cleanup attempts {{(pid=63088) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1117.154833] env[63088]: DEBUG nova.objects.instance [None req-60ee2949-6f9b-4a40-8984-0f26595d38e8 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lazy-loading 'migration_context' on Instance uuid b25b5883-20c0-4cc9-8cc1-2c22a0f5be81 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1117.301620] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1117.301803] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Cleaning up deleted instances with incomplete migration {{(pid=63088) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11264}} [ 1117.450942] env[63088]: DEBUG oslo_concurrency.lockutils [None req-af441786-2d51-4cbf-ab87-ac35d2d4a26e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquiring lock "refresh_cache-2b87a9f1-5f10-43c2-8bc9-6d560ec88015" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1117.451174] env[63088]: DEBUG oslo_concurrency.lockutils [None req-af441786-2d51-4cbf-ab87-ac35d2d4a26e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquired lock "refresh_cache-2b87a9f1-5f10-43c2-8bc9-6d560ec88015" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1117.451358] env[63088]: DEBUG nova.network.neutron [None req-af441786-2d51-4cbf-ab87-ac35d2d4a26e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1117.735951] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1942bdce-b6fe-408d-9975-65b764ea022b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.744413] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43423f27-df43-4d79-b0b4-0852d07e4c90 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.774863] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f6a02d6-ffe6-4699-a0a6-ab621e09a3f2 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.782307] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa291694-ceac-4187-9b48-070e73a77dae {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.795777] env[63088]: DEBUG nova.compute.provider_tree [None req-60ee2949-6f9b-4a40-8984-0f26595d38e8 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1117.804485] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1118.172618] env[63088]: DEBUG nova.network.neutron [None req-af441786-2d51-4cbf-ab87-ac35d2d4a26e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Updating instance_info_cache with network_info: [{"id": "f88388b7-bc70-4095-a792-8d155ef34534", "address": "fa:16:3e:f7:d1:c7", "network": {"id": "711d3d9a-86db-4aa1-87b5-481f1812dfa4", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-59376186-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.225", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a7898ddafe0d41038a6ae4277f2c2f48", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e41070eb-3ac1-4ca9-a3d0-fd65893a97de", "external-id": "nsx-vlan-transportzone-596", "segmentation_id": 596, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf88388b7-bc", "ovs_interfaceid": "f88388b7-bc70-4095-a792-8d155ef34534", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1118.298918] env[63088]: DEBUG nova.scheduler.client.report [None req-60ee2949-6f9b-4a40-8984-0f26595d38e8 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1118.675194] env[63088]: DEBUG oslo_concurrency.lockutils [None req-af441786-2d51-4cbf-ab87-ac35d2d4a26e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Releasing lock "refresh_cache-2b87a9f1-5f10-43c2-8bc9-6d560ec88015" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1119.196569] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3586d553-a099-4eb1-b6af-0c6545b01c93 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.216150] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ced859dd-aa90-4f61-97c0-38e3d54a8baa {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.223344] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-af441786-2d51-4cbf-ab87-ac35d2d4a26e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Updating instance '2b87a9f1-5f10-43c2-8bc9-6d560ec88015' progress to 83 {{(pid=63088) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1119.309065] env[63088]: DEBUG oslo_concurrency.lockutils [None req-60ee2949-6f9b-4a40-8984-0f26595d38e8 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.657s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1119.730175] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-af441786-2d51-4cbf-ab87-ac35d2d4a26e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Updating instance '2b87a9f1-5f10-43c2-8bc9-6d560ec88015' progress to 100 {{(pid=63088) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1120.306249] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1120.306620] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1120.811502] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1120.811687] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Starting heal instance info cache {{(pid=63088) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1120.843108] env[63088]: INFO nova.compute.manager [None req-60ee2949-6f9b-4a40-8984-0f26595d38e8 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Swapping old allocation on dict_keys(['6eae54a9-8831-40eb-bf54-4bc60d346b02']) held by migration e77e7800-a0d0-4286-bebd-ba1e4a27bf70 for instance [ 1120.865874] env[63088]: DEBUG nova.scheduler.client.report [None req-60ee2949-6f9b-4a40-8984-0f26595d38e8 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Overwriting current allocation {'allocations': {'6eae54a9-8831-40eb-bf54-4bc60d346b02': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 133}}, 'project_id': 'dace8b5181b84623b08f903d12dfd31e', 'user_id': '787d7bd9f8e7490c9f24ff0e159bc98f', 'consumer_generation': 1} on consumer b25b5883-20c0-4cc9-8cc1-2c22a0f5be81 {{(pid=63088) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 1120.946419] env[63088]: DEBUG oslo_concurrency.lockutils [None req-60ee2949-6f9b-4a40-8984-0f26595d38e8 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquiring lock "refresh_cache-b25b5883-20c0-4cc9-8cc1-2c22a0f5be81" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1120.946677] env[63088]: DEBUG oslo_concurrency.lockutils [None req-60ee2949-6f9b-4a40-8984-0f26595d38e8 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquired lock "refresh_cache-b25b5883-20c0-4cc9-8cc1-2c22a0f5be81" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1120.946855] env[63088]: DEBUG nova.network.neutron [None req-60ee2949-6f9b-4a40-8984-0f26595d38e8 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1121.342048] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Acquiring lock "refresh_cache-2b87a9f1-5f10-43c2-8bc9-6d560ec88015" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1121.342292] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Acquired lock "refresh_cache-2b87a9f1-5f10-43c2-8bc9-6d560ec88015" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1121.342376] env[63088]: DEBUG nova.network.neutron [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Forcefully refreshing network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1121.689662] env[63088]: DEBUG nova.network.neutron [None req-60ee2949-6f9b-4a40-8984-0f26595d38e8 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Updating instance_info_cache with network_info: [{"id": "f344e65c-4141-4c1d-a8e0-4aa1571a8e2f", "address": "fa:16:3e:3f:78:a4", "network": {"id": "dfcbe153-bc01-4362-9247-de9f0b1c847b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-841757397-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.152", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dace8b5181b84623b08f903d12dfd31e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c68b7663-4f0e-47f0-ac7f-40c6d952f7bb", "external-id": "nsx-vlan-transportzone-696", "segmentation_id": 696, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf344e65c-41", "ovs_interfaceid": "f344e65c-4141-4c1d-a8e0-4aa1571a8e2f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1121.870040] env[63088]: DEBUG oslo_concurrency.lockutils [None req-44dedbce-2808-40e5-8db7-7a5980f055c6 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquiring lock "2b87a9f1-5f10-43c2-8bc9-6d560ec88015" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1121.870234] env[63088]: DEBUG oslo_concurrency.lockutils [None req-44dedbce-2808-40e5-8db7-7a5980f055c6 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lock "2b87a9f1-5f10-43c2-8bc9-6d560ec88015" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1121.871025] env[63088]: DEBUG nova.compute.manager [None req-44dedbce-2808-40e5-8db7-7a5980f055c6 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Going to confirm migration 7 {{(pid=63088) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1122.192798] env[63088]: DEBUG oslo_concurrency.lockutils [None req-60ee2949-6f9b-4a40-8984-0f26595d38e8 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Releasing lock "refresh_cache-b25b5883-20c0-4cc9-8cc1-2c22a0f5be81" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1122.193319] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-60ee2949-6f9b-4a40-8984-0f26595d38e8 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1122.193624] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0e94121b-36fc-45fe-b940-168c3c1edb87 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.200808] env[63088]: DEBUG oslo_vmware.api [None req-60ee2949-6f9b-4a40-8984-0f26595d38e8 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for the task: (returnval){ [ 1122.200808] env[63088]: value = "task-1285548" [ 1122.200808] env[63088]: _type = "Task" [ 1122.200808] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1122.209172] env[63088]: DEBUG oslo_vmware.api [None req-60ee2949-6f9b-4a40-8984-0f26595d38e8 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285548, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.425595] env[63088]: DEBUG oslo_concurrency.lockutils [None req-44dedbce-2808-40e5-8db7-7a5980f055c6 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquiring lock "refresh_cache-2b87a9f1-5f10-43c2-8bc9-6d560ec88015" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1122.629390] env[63088]: DEBUG nova.network.neutron [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Updating instance_info_cache with network_info: [{"id": "f88388b7-bc70-4095-a792-8d155ef34534", "address": "fa:16:3e:f7:d1:c7", "network": {"id": "711d3d9a-86db-4aa1-87b5-481f1812dfa4", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-59376186-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.225", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a7898ddafe0d41038a6ae4277f2c2f48", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e41070eb-3ac1-4ca9-a3d0-fd65893a97de", "external-id": "nsx-vlan-transportzone-596", "segmentation_id": 596, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf88388b7-bc", "ovs_interfaceid": "f88388b7-bc70-4095-a792-8d155ef34534", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1122.711672] env[63088]: DEBUG oslo_vmware.api [None req-60ee2949-6f9b-4a40-8984-0f26595d38e8 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285548, 'name': PowerOffVM_Task, 'duration_secs': 0.22339} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1122.712019] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-60ee2949-6f9b-4a40-8984-0f26595d38e8 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1122.712898] env[63088]: DEBUG nova.virt.hardware [None req-60ee2949-6f9b-4a40-8984-0f26595d38e8 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1122.713145] env[63088]: DEBUG nova.virt.hardware [None req-60ee2949-6f9b-4a40-8984-0f26595d38e8 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1122.713310] env[63088]: DEBUG nova.virt.hardware [None req-60ee2949-6f9b-4a40-8984-0f26595d38e8 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1122.713498] env[63088]: DEBUG nova.virt.hardware [None req-60ee2949-6f9b-4a40-8984-0f26595d38e8 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1122.713645] env[63088]: DEBUG nova.virt.hardware [None req-60ee2949-6f9b-4a40-8984-0f26595d38e8 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1122.713803] env[63088]: DEBUG nova.virt.hardware [None req-60ee2949-6f9b-4a40-8984-0f26595d38e8 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1122.714013] env[63088]: DEBUG nova.virt.hardware [None req-60ee2949-6f9b-4a40-8984-0f26595d38e8 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1122.714184] env[63088]: DEBUG nova.virt.hardware [None req-60ee2949-6f9b-4a40-8984-0f26595d38e8 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1122.714353] env[63088]: DEBUG nova.virt.hardware [None req-60ee2949-6f9b-4a40-8984-0f26595d38e8 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1122.714519] env[63088]: DEBUG nova.virt.hardware [None req-60ee2949-6f9b-4a40-8984-0f26595d38e8 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1122.714694] env[63088]: DEBUG nova.virt.hardware [None req-60ee2949-6f9b-4a40-8984-0f26595d38e8 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1122.719670] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c51ced68-b66e-43ed-a1df-e9a0204947e8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.735014] env[63088]: DEBUG oslo_vmware.api [None req-60ee2949-6f9b-4a40-8984-0f26595d38e8 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for the task: (returnval){ [ 1122.735014] env[63088]: value = "task-1285549" [ 1122.735014] env[63088]: _type = "Task" [ 1122.735014] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1122.743032] env[63088]: DEBUG oslo_vmware.api [None req-60ee2949-6f9b-4a40-8984-0f26595d38e8 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285549, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.131878] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Releasing lock "refresh_cache-2b87a9f1-5f10-43c2-8bc9-6d560ec88015" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1123.132183] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Updated the network info_cache for instance {{(pid=63088) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 1123.132536] env[63088]: DEBUG oslo_concurrency.lockutils [None req-44dedbce-2808-40e5-8db7-7a5980f055c6 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquired lock "refresh_cache-2b87a9f1-5f10-43c2-8bc9-6d560ec88015" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1123.132746] env[63088]: DEBUG nova.network.neutron [None req-44dedbce-2808-40e5-8db7-7a5980f055c6 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1123.132976] env[63088]: DEBUG nova.objects.instance [None req-44dedbce-2808-40e5-8db7-7a5980f055c6 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lazy-loading 'info_cache' on Instance uuid 2b87a9f1-5f10-43c2-8bc9-6d560ec88015 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1123.134368] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1123.134567] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1123.134986] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1123.135168] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1123.135319] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1123.135466] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1123.135597] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63088) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1123.135742] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager.update_available_resource {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1123.245430] env[63088]: DEBUG oslo_vmware.api [None req-60ee2949-6f9b-4a40-8984-0f26595d38e8 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285549, 'name': ReconfigVM_Task, 'duration_secs': 0.14405} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1123.246080] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2f8ac88-86de-4d7e-8617-374a94eed9a8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.266163] env[63088]: DEBUG nova.virt.hardware [None req-60ee2949-6f9b-4a40-8984-0f26595d38e8 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1123.266444] env[63088]: DEBUG nova.virt.hardware [None req-60ee2949-6f9b-4a40-8984-0f26595d38e8 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1123.266613] env[63088]: DEBUG nova.virt.hardware [None req-60ee2949-6f9b-4a40-8984-0f26595d38e8 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1123.266803] env[63088]: DEBUG nova.virt.hardware [None req-60ee2949-6f9b-4a40-8984-0f26595d38e8 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1123.266956] env[63088]: DEBUG nova.virt.hardware [None req-60ee2949-6f9b-4a40-8984-0f26595d38e8 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1123.267142] env[63088]: DEBUG nova.virt.hardware [None req-60ee2949-6f9b-4a40-8984-0f26595d38e8 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1123.267356] env[63088]: DEBUG nova.virt.hardware [None req-60ee2949-6f9b-4a40-8984-0f26595d38e8 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1123.267525] env[63088]: DEBUG nova.virt.hardware [None req-60ee2949-6f9b-4a40-8984-0f26595d38e8 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1123.267692] env[63088]: DEBUG nova.virt.hardware [None req-60ee2949-6f9b-4a40-8984-0f26595d38e8 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1123.267859] env[63088]: DEBUG nova.virt.hardware [None req-60ee2949-6f9b-4a40-8984-0f26595d38e8 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1123.268049] env[63088]: DEBUG nova.virt.hardware [None req-60ee2949-6f9b-4a40-8984-0f26595d38e8 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1123.269321] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8ce924be-11fc-491a-928e-6abea1b20a95 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.274526] env[63088]: DEBUG oslo_vmware.api [None req-60ee2949-6f9b-4a40-8984-0f26595d38e8 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for the task: (returnval){ [ 1123.274526] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52259cec-1246-b91f-58d0-3af6a2e5ec51" [ 1123.274526] env[63088]: _type = "Task" [ 1123.274526] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1123.282810] env[63088]: DEBUG oslo_vmware.api [None req-60ee2949-6f9b-4a40-8984-0f26595d38e8 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52259cec-1246-b91f-58d0-3af6a2e5ec51, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.638618] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1123.638618] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1123.639031] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1123.639031] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63088) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1123.639975] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2132072-b56c-48b0-bf69-5826e8fa5ac1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.648237] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1124369-7375-442d-b792-65102c3b1f34 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.662286] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6eb0b8a-6f18-40fd-9105-d407ca53f13b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.668851] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ce2c5a3-fb4c-41ae-a667-5ab89f8e86a8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.698599] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180856MB free_disk=140GB free_vcpus=48 pci_devices=None {{(pid=63088) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1123.698761] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1123.699013] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1123.784053] env[63088]: DEBUG oslo_vmware.api [None req-60ee2949-6f9b-4a40-8984-0f26595d38e8 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52259cec-1246-b91f-58d0-3af6a2e5ec51, 'name': SearchDatastore_Task, 'duration_secs': 0.009941} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1123.789339] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-60ee2949-6f9b-4a40-8984-0f26595d38e8 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Reconfiguring VM instance instance-00000067 to detach disk 2000 {{(pid=63088) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1123.789921] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2d42aee8-9498-4ad0-a627-3b67b2a7ca41 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.807549] env[63088]: DEBUG oslo_vmware.api [None req-60ee2949-6f9b-4a40-8984-0f26595d38e8 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for the task: (returnval){ [ 1123.807549] env[63088]: value = "task-1285550" [ 1123.807549] env[63088]: _type = "Task" [ 1123.807549] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1123.815308] env[63088]: DEBUG oslo_vmware.api [None req-60ee2949-6f9b-4a40-8984-0f26595d38e8 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285550, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.318179] env[63088]: DEBUG oslo_vmware.api [None req-60ee2949-6f9b-4a40-8984-0f26595d38e8 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285550, 'name': ReconfigVM_Task, 'duration_secs': 0.224569} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1124.318179] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-60ee2949-6f9b-4a40-8984-0f26595d38e8 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Reconfigured VM instance instance-00000067 to detach disk 2000 {{(pid=63088) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1124.318395] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2dc2c86-7952-4f2f-b2a2-eb5f4a909771 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.339770] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-60ee2949-6f9b-4a40-8984-0f26595d38e8 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Reconfiguring VM instance instance-00000067 to attach disk [datastore2] b25b5883-20c0-4cc9-8cc1-2c22a0f5be81/b25b5883-20c0-4cc9-8cc1-2c22a0f5be81.vmdk or device None with type thin {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1124.342301] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d8823818-ec3d-4704-8702-b74dd7d3ab2e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.360769] env[63088]: DEBUG oslo_vmware.api [None req-60ee2949-6f9b-4a40-8984-0f26595d38e8 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for the task: (returnval){ [ 1124.360769] env[63088]: value = "task-1285551" [ 1124.360769] env[63088]: _type = "Task" [ 1124.360769] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.369354] env[63088]: DEBUG oslo_vmware.api [None req-60ee2949-6f9b-4a40-8984-0f26595d38e8 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285551, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.391335] env[63088]: DEBUG nova.network.neutron [None req-44dedbce-2808-40e5-8db7-7a5980f055c6 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Updating instance_info_cache with network_info: [{"id": "f88388b7-bc70-4095-a792-8d155ef34534", "address": "fa:16:3e:f7:d1:c7", "network": {"id": "711d3d9a-86db-4aa1-87b5-481f1812dfa4", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-59376186-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.225", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a7898ddafe0d41038a6ae4277f2c2f48", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e41070eb-3ac1-4ca9-a3d0-fd65893a97de", "external-id": "nsx-vlan-transportzone-596", "segmentation_id": 596, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf88388b7-bc", "ovs_interfaceid": "f88388b7-bc70-4095-a792-8d155ef34534", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1124.706310] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Applying migration context for instance 2b87a9f1-5f10-43c2-8bc9-6d560ec88015 as it has an incoming, in-progress migration 74ec533f-3231-45f3-93f2-fa9c6481ffc0. Migration status is confirming {{(pid=63088) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1124.707078] env[63088]: INFO nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Updating resource usage from migration 74ec533f-3231-45f3-93f2-fa9c6481ffc0 [ 1124.725191] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance 3fa11db4-2f70-495e-a562-ac5f06153847 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1124.725375] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance b25b5883-20c0-4cc9-8cc1-2c22a0f5be81 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1124.725500] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Migration 74ec533f-3231-45f3-93f2-fa9c6481ffc0 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1124.725617] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance 2b87a9f1-5f10-43c2-8bc9-6d560ec88015 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1124.725809] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=63088) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1124.725948] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1344MB phys_disk=200GB used_disk=4GB total_vcpus=48 used_vcpus=4 pci_stats=[] {{(pid=63088) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1124.786215] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db3de987-fea3-4366-aa53-8f1bceb41229 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.794238] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a68000d4-7be3-4ae6-aee1-301c2f20c8e3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.823684] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dff125c-4eb0-4153-b8b7-0fb96b4b4d39 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.831427] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-637153a8-f3b9-4450-83e0-bfcce4d42866 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.845339] env[63088]: DEBUG nova.compute.provider_tree [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1124.869048] env[63088]: DEBUG oslo_vmware.api [None req-60ee2949-6f9b-4a40-8984-0f26595d38e8 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285551, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.893089] env[63088]: DEBUG oslo_vmware.rw_handles [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c35194-26f1-f149-0043-900414536f3c/disk-0.vmdk. {{(pid=63088) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1124.893636] env[63088]: DEBUG oslo_concurrency.lockutils [None req-44dedbce-2808-40e5-8db7-7a5980f055c6 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Releasing lock "refresh_cache-2b87a9f1-5f10-43c2-8bc9-6d560ec88015" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1124.893904] env[63088]: DEBUG nova.objects.instance [None req-44dedbce-2808-40e5-8db7-7a5980f055c6 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lazy-loading 'migration_context' on Instance uuid 2b87a9f1-5f10-43c2-8bc9-6d560ec88015 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1124.895449] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3554545b-cfdb-4175-840d-f86745d12b67 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.901286] env[63088]: DEBUG oslo_vmware.rw_handles [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c35194-26f1-f149-0043-900414536f3c/disk-0.vmdk is in state: ready. {{(pid=63088) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1124.901453] env[63088]: ERROR oslo_vmware.rw_handles [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c35194-26f1-f149-0043-900414536f3c/disk-0.vmdk due to incomplete transfer. [ 1124.902185] env[63088]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-2c649c02-16c4-4f19-a88a-4f6b197d09df {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.908847] env[63088]: DEBUG oslo_vmware.rw_handles [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c35194-26f1-f149-0043-900414536f3c/disk-0.vmdk. {{(pid=63088) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1124.909047] env[63088]: DEBUG nova.virt.vmwareapi.images [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Uploaded image f1db6e74-0b10-490d-a958-6fa032fe3caf to the Glance image server {{(pid=63088) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1124.911361] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Destroying the VM {{(pid=63088) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1124.911789] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-472245c1-13d6-454d-ae75-6b46a228ead5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.917212] env[63088]: DEBUG oslo_vmware.api [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Waiting for the task: (returnval){ [ 1124.917212] env[63088]: value = "task-1285552" [ 1124.917212] env[63088]: _type = "Task" [ 1124.917212] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.925339] env[63088]: DEBUG oslo_vmware.api [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285552, 'name': Destroy_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.348330] env[63088]: DEBUG nova.scheduler.client.report [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1125.371739] env[63088]: DEBUG oslo_vmware.api [None req-60ee2949-6f9b-4a40-8984-0f26595d38e8 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285551, 'name': ReconfigVM_Task, 'duration_secs': 0.652525} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.372036] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-60ee2949-6f9b-4a40-8984-0f26595d38e8 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Reconfigured VM instance instance-00000067 to attach disk [datastore2] b25b5883-20c0-4cc9-8cc1-2c22a0f5be81/b25b5883-20c0-4cc9-8cc1-2c22a0f5be81.vmdk or device None with type thin {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1125.372868] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce9cb03e-3b4d-4894-9332-9c801145cc94 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.390886] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23b9c30f-f776-40a3-9787-568c29813132 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.408050] env[63088]: DEBUG nova.objects.base [None req-44dedbce-2808-40e5-8db7-7a5980f055c6 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Object Instance<2b87a9f1-5f10-43c2-8bc9-6d560ec88015> lazy-loaded attributes: info_cache,migration_context {{(pid=63088) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1125.408948] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b282ae4-d7c4-4663-9b3d-207d9c2b6fa5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.411664] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff1be6b4-601d-4f91-9d9e-4cbcd515c807 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.446668] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ce57efe4-8264-47d5-82f7-0ab46dfadcb8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.450156] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc0a65e1-98f4-484e-8983-69e304195bab {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.456282] env[63088]: DEBUG oslo_vmware.api [None req-44dedbce-2808-40e5-8db7-7a5980f055c6 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1125.456282] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52071813-cb9d-878f-978d-d80c9472ec6d" [ 1125.456282] env[63088]: _type = "Task" [ 1125.456282] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.461070] env[63088]: DEBUG oslo_vmware.api [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285552, 'name': Destroy_Task, 'duration_secs': 0.317533} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.461415] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-60ee2949-6f9b-4a40-8984-0f26595d38e8 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1125.464220] env[63088]: INFO nova.virt.vmwareapi.vm_util [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Destroyed the VM [ 1125.464464] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Deleting Snapshot of the VM instance {{(pid=63088) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1125.464686] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-03cd681e-3b86-4a5e-b18b-f95fdab0f08d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.466009] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-c4e6cbb0-93f0-4204-84a2-aa885f6bb06e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.472749] env[63088]: DEBUG oslo_vmware.api [None req-44dedbce-2808-40e5-8db7-7a5980f055c6 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52071813-cb9d-878f-978d-d80c9472ec6d, 'name': SearchDatastore_Task, 'duration_secs': 0.006776} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.474637] env[63088]: DEBUG oslo_concurrency.lockutils [None req-44dedbce-2808-40e5-8db7-7a5980f055c6 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1125.474949] env[63088]: DEBUG oslo_vmware.api [None req-60ee2949-6f9b-4a40-8984-0f26595d38e8 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for the task: (returnval){ [ 1125.474949] env[63088]: value = "task-1285553" [ 1125.474949] env[63088]: _type = "Task" [ 1125.474949] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.475199] env[63088]: DEBUG oslo_vmware.api [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Waiting for the task: (returnval){ [ 1125.475199] env[63088]: value = "task-1285554" [ 1125.475199] env[63088]: _type = "Task" [ 1125.475199] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.485828] env[63088]: DEBUG oslo_vmware.api [None req-60ee2949-6f9b-4a40-8984-0f26595d38e8 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285553, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.488573] env[63088]: DEBUG oslo_vmware.api [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285554, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.853632] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63088) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1125.853982] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.155s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1125.854193] env[63088]: DEBUG oslo_concurrency.lockutils [None req-44dedbce-2808-40e5-8db7-7a5980f055c6 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.380s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1125.987714] env[63088]: DEBUG oslo_vmware.api [None req-60ee2949-6f9b-4a40-8984-0f26595d38e8 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285553, 'name': PowerOnVM_Task, 'duration_secs': 0.372959} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.990541] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-60ee2949-6f9b-4a40-8984-0f26595d38e8 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1125.993789] env[63088]: DEBUG oslo_vmware.api [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285554, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.410421] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8e856d8-24bb-4edb-8210-95d7abcfd999 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.417790] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9444e99b-08eb-4bff-9a4e-2c2cef2f3248 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.446316] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf83c937-c155-4896-b420-6d6aca4e2d33 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.452960] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7de35e53-ec3e-4881-b5f1-06f5a87232c9 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.466618] env[63088]: DEBUG nova.compute.provider_tree [None req-44dedbce-2808-40e5-8db7-7a5980f055c6 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1126.484505] env[63088]: DEBUG oslo_vmware.api [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285554, 'name': RemoveSnapshot_Task, 'duration_secs': 0.758412} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.484749] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Deleted Snapshot of the VM instance {{(pid=63088) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1126.485074] env[63088]: DEBUG nova.compute.manager [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1126.485775] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac7c5a35-c052-4cbe-9208-5cd7cb76be67 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.969585] env[63088]: DEBUG nova.scheduler.client.report [None req-44dedbce-2808-40e5-8db7-7a5980f055c6 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1126.998023] env[63088]: INFO nova.compute.manager [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Shelve offloading [ 1126.999461] env[63088]: INFO nova.compute.manager [None req-60ee2949-6f9b-4a40-8984-0f26595d38e8 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Updating instance to original state: 'active' [ 1127.002514] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1127.002763] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-998e5207-5153-4a0c-b3d0-1b023b1cc9b4 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.010646] env[63088]: DEBUG oslo_vmware.api [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Waiting for the task: (returnval){ [ 1127.010646] env[63088]: value = "task-1285555" [ 1127.010646] env[63088]: _type = "Task" [ 1127.010646] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.020497] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] VM already powered off {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1127.020768] env[63088]: DEBUG nova.compute.manager [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1127.021512] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eff9408f-70a1-4df7-b17b-ab2b066d2531 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.026960] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Acquiring lock "refresh_cache-3fa11db4-2f70-495e-a562-ac5f06153847" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1127.027144] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Acquired lock "refresh_cache-3fa11db4-2f70-495e-a562-ac5f06153847" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1127.027318] env[63088]: DEBUG nova.network.neutron [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1127.783928] env[63088]: DEBUG nova.network.neutron [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Updating instance_info_cache with network_info: [{"id": "eb4d6233-2d9d-4756-8b9d-084bdfaeb3f7", "address": "fa:16:3e:c2:db:1d", "network": {"id": "cc2c0304-8906-443a-a76e-e7a32158346f", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1867758042-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.195", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b18ddddf2314d83addf550b8cb91977", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0f096917-a0cf-4add-a9d2-23ca1c723b3b", "external-id": "nsx-vlan-transportzone-894", "segmentation_id": 894, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeb4d6233-2d", "ovs_interfaceid": "eb4d6233-2d9d-4756-8b9d-084bdfaeb3f7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1127.979833] env[63088]: DEBUG oslo_concurrency.lockutils [None req-44dedbce-2808-40e5-8db7-7a5980f055c6 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.126s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1127.980089] env[63088]: DEBUG nova.compute.manager [None req-44dedbce-2808-40e5-8db7-7a5980f055c6 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Resized/migrated instance is powered off. Setting vm_state to 'stopped'. {{(pid=63088) _confirm_resize /opt/stack/nova/nova/compute/manager.py:4909}} [ 1128.286687] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Releasing lock "refresh_cache-3fa11db4-2f70-495e-a562-ac5f06153847" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1128.506210] env[63088]: DEBUG nova.compute.manager [req-e1e678b1-0bfb-4cea-be32-07321d6ee2a3 req-c920b48c-8d76-400e-9523-9ee70c6406a3 service nova] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Received event network-vif-unplugged-eb4d6233-2d9d-4756-8b9d-084bdfaeb3f7 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1128.506439] env[63088]: DEBUG oslo_concurrency.lockutils [req-e1e678b1-0bfb-4cea-be32-07321d6ee2a3 req-c920b48c-8d76-400e-9523-9ee70c6406a3 service nova] Acquiring lock "3fa11db4-2f70-495e-a562-ac5f06153847-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1128.506652] env[63088]: DEBUG oslo_concurrency.lockutils [req-e1e678b1-0bfb-4cea-be32-07321d6ee2a3 req-c920b48c-8d76-400e-9523-9ee70c6406a3 service nova] Lock "3fa11db4-2f70-495e-a562-ac5f06153847-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1128.506823] env[63088]: DEBUG oslo_concurrency.lockutils [req-e1e678b1-0bfb-4cea-be32-07321d6ee2a3 req-c920b48c-8d76-400e-9523-9ee70c6406a3 service nova] Lock "3fa11db4-2f70-495e-a562-ac5f06153847-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1128.507047] env[63088]: DEBUG nova.compute.manager [req-e1e678b1-0bfb-4cea-be32-07321d6ee2a3 req-c920b48c-8d76-400e-9523-9ee70c6406a3 service nova] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] No waiting events found dispatching network-vif-unplugged-eb4d6233-2d9d-4756-8b9d-084bdfaeb3f7 {{(pid=63088) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1128.507178] env[63088]: WARNING nova.compute.manager [req-e1e678b1-0bfb-4cea-be32-07321d6ee2a3 req-c920b48c-8d76-400e-9523-9ee70c6406a3 service nova] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Received unexpected event network-vif-unplugged-eb4d6233-2d9d-4756-8b9d-084bdfaeb3f7 for instance with vm_state shelved and task_state shelving_offloading. [ 1128.533976] env[63088]: DEBUG oslo_concurrency.lockutils [None req-7424b985-8d50-4561-b4cf-cb870c1731d5 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquiring lock "b25b5883-20c0-4cc9-8cc1-2c22a0f5be81" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1128.534274] env[63088]: DEBUG oslo_concurrency.lockutils [None req-7424b985-8d50-4561-b4cf-cb870c1731d5 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lock "b25b5883-20c0-4cc9-8cc1-2c22a0f5be81" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1128.534536] env[63088]: DEBUG oslo_concurrency.lockutils [None req-7424b985-8d50-4561-b4cf-cb870c1731d5 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquiring lock "b25b5883-20c0-4cc9-8cc1-2c22a0f5be81-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1128.534736] env[63088]: DEBUG oslo_concurrency.lockutils [None req-7424b985-8d50-4561-b4cf-cb870c1731d5 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lock "b25b5883-20c0-4cc9-8cc1-2c22a0f5be81-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1128.534940] env[63088]: DEBUG oslo_concurrency.lockutils [None req-7424b985-8d50-4561-b4cf-cb870c1731d5 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lock "b25b5883-20c0-4cc9-8cc1-2c22a0f5be81-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1128.536978] env[63088]: INFO nova.compute.manager [None req-7424b985-8d50-4561-b4cf-cb870c1731d5 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Terminating instance [ 1128.539084] env[63088]: DEBUG nova.compute.manager [None req-7424b985-8d50-4561-b4cf-cb870c1731d5 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1128.539290] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-7424b985-8d50-4561-b4cf-cb870c1731d5 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1128.540476] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e647a2a-6dac-4085-9b01-367698213875 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.545759] env[63088]: INFO nova.scheduler.client.report [None req-44dedbce-2808-40e5-8db7-7a5980f055c6 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Deleted allocation for migration 74ec533f-3231-45f3-93f2-fa9c6481ffc0 [ 1128.551939] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-7424b985-8d50-4561-b4cf-cb870c1731d5 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1128.552408] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3e9149d0-43c8-419a-9e7a-930f5d12004a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.560169] env[63088]: DEBUG oslo_vmware.api [None req-7424b985-8d50-4561-b4cf-cb870c1731d5 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for the task: (returnval){ [ 1128.560169] env[63088]: value = "task-1285556" [ 1128.560169] env[63088]: _type = "Task" [ 1128.560169] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.568036] env[63088]: DEBUG oslo_vmware.api [None req-7424b985-8d50-4561-b4cf-cb870c1731d5 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285556, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.605695] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1128.606928] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cafc6386-5cd7-4a54-83d6-53445dcf9e1f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.616128] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Unregistering the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1128.616408] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f1cfe4c3-c37d-4855-b9e2-8df205336ce3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.684265] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Unregistered the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1128.684532] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Deleting contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1128.684739] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Deleting the datastore file [datastore1] 3fa11db4-2f70-495e-a562-ac5f06153847 {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1128.685069] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ffc0c1f5-e97b-4b1a-8ec2-8d63160df84c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.691815] env[63088]: DEBUG oslo_vmware.api [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Waiting for the task: (returnval){ [ 1128.691815] env[63088]: value = "task-1285558" [ 1128.691815] env[63088]: _type = "Task" [ 1128.691815] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.699526] env[63088]: DEBUG oslo_vmware.api [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285558, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.794784] env[63088]: DEBUG nova.objects.instance [None req-91152bfa-3f14-496a-9e3c-24896547bba0 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lazy-loading 'flavor' on Instance uuid 2b87a9f1-5f10-43c2-8bc9-6d560ec88015 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1129.054383] env[63088]: DEBUG oslo_concurrency.lockutils [None req-44dedbce-2808-40e5-8db7-7a5980f055c6 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lock "2b87a9f1-5f10-43c2-8bc9-6d560ec88015" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 7.184s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1129.070127] env[63088]: DEBUG oslo_vmware.api [None req-7424b985-8d50-4561-b4cf-cb870c1731d5 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285556, 'name': PowerOffVM_Task, 'duration_secs': 0.199217} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.070467] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-7424b985-8d50-4561-b4cf-cb870c1731d5 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1129.070718] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-7424b985-8d50-4561-b4cf-cb870c1731d5 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Unregistering the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1129.071024] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-de3ec811-4ab5-44e5-8878-21a2484024e2 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.132267] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-7424b985-8d50-4561-b4cf-cb870c1731d5 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Unregistered the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1129.132588] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-7424b985-8d50-4561-b4cf-cb870c1731d5 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Deleting contents of the VM from datastore datastore2 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1129.132913] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-7424b985-8d50-4561-b4cf-cb870c1731d5 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Deleting the datastore file [datastore2] b25b5883-20c0-4cc9-8cc1-2c22a0f5be81 {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1129.133246] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-dd788e33-0e69-4edf-837a-af72bd529ad5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.139790] env[63088]: DEBUG oslo_vmware.api [None req-7424b985-8d50-4561-b4cf-cb870c1731d5 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for the task: (returnval){ [ 1129.139790] env[63088]: value = "task-1285560" [ 1129.139790] env[63088]: _type = "Task" [ 1129.139790] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.147366] env[63088]: DEBUG oslo_vmware.api [None req-7424b985-8d50-4561-b4cf-cb870c1731d5 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285560, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.201102] env[63088]: DEBUG oslo_vmware.api [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285558, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.183087} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.201349] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1129.201539] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Deleted contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1129.201720] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1129.218910] env[63088]: INFO nova.scheduler.client.report [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Deleted allocations for instance 3fa11db4-2f70-495e-a562-ac5f06153847 [ 1129.299845] env[63088]: DEBUG oslo_concurrency.lockutils [None req-91152bfa-3f14-496a-9e3c-24896547bba0 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquiring lock "refresh_cache-2b87a9f1-5f10-43c2-8bc9-6d560ec88015" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1129.300071] env[63088]: DEBUG oslo_concurrency.lockutils [None req-91152bfa-3f14-496a-9e3c-24896547bba0 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquired lock "refresh_cache-2b87a9f1-5f10-43c2-8bc9-6d560ec88015" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1129.300255] env[63088]: DEBUG nova.network.neutron [None req-91152bfa-3f14-496a-9e3c-24896547bba0 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1129.300431] env[63088]: DEBUG nova.objects.instance [None req-91152bfa-3f14-496a-9e3c-24896547bba0 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lazy-loading 'info_cache' on Instance uuid 2b87a9f1-5f10-43c2-8bc9-6d560ec88015 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1129.649144] env[63088]: DEBUG oslo_vmware.api [None req-7424b985-8d50-4561-b4cf-cb870c1731d5 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285560, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.151596} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.649407] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-7424b985-8d50-4561-b4cf-cb870c1731d5 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1129.649593] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-7424b985-8d50-4561-b4cf-cb870c1731d5 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Deleted contents of the VM from datastore datastore2 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1129.649775] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-7424b985-8d50-4561-b4cf-cb870c1731d5 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1129.649950] env[63088]: INFO nova.compute.manager [None req-7424b985-8d50-4561-b4cf-cb870c1731d5 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1129.650211] env[63088]: DEBUG oslo.service.loopingcall [None req-7424b985-8d50-4561-b4cf-cb870c1731d5 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1129.650408] env[63088]: DEBUG nova.compute.manager [-] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1129.650504] env[63088]: DEBUG nova.network.neutron [-] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1129.723240] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1129.723510] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1129.723739] env[63088]: DEBUG nova.objects.instance [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Lazy-loading 'resources' on Instance uuid 3fa11db4-2f70-495e-a562-ac5f06153847 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1129.803540] env[63088]: DEBUG nova.objects.base [None req-91152bfa-3f14-496a-9e3c-24896547bba0 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Object Instance<2b87a9f1-5f10-43c2-8bc9-6d560ec88015> lazy-loaded attributes: flavor,info_cache {{(pid=63088) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1130.226345] env[63088]: DEBUG nova.objects.instance [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Lazy-loading 'numa_topology' on Instance uuid 3fa11db4-2f70-495e-a562-ac5f06153847 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1130.507859] env[63088]: DEBUG nova.network.neutron [None req-91152bfa-3f14-496a-9e3c-24896547bba0 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Updating instance_info_cache with network_info: [{"id": "f88388b7-bc70-4095-a792-8d155ef34534", "address": "fa:16:3e:f7:d1:c7", "network": {"id": "711d3d9a-86db-4aa1-87b5-481f1812dfa4", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-59376186-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.225", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a7898ddafe0d41038a6ae4277f2c2f48", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e41070eb-3ac1-4ca9-a3d0-fd65893a97de", "external-id": "nsx-vlan-transportzone-596", "segmentation_id": 596, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf88388b7-bc", "ovs_interfaceid": "f88388b7-bc70-4095-a792-8d155ef34534", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1130.530464] env[63088]: DEBUG nova.network.neutron [-] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1130.592748] env[63088]: DEBUG nova.compute.manager [req-47a51e1f-8c27-4d36-a9c8-b1053c46f395 req-1563ee9b-6031-4e63-8d89-a431425f208e service nova] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Received event network-changed-eb4d6233-2d9d-4756-8b9d-084bdfaeb3f7 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1130.593087] env[63088]: DEBUG nova.compute.manager [req-47a51e1f-8c27-4d36-a9c8-b1053c46f395 req-1563ee9b-6031-4e63-8d89-a431425f208e service nova] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Refreshing instance network info cache due to event network-changed-eb4d6233-2d9d-4756-8b9d-084bdfaeb3f7. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1130.593380] env[63088]: DEBUG oslo_concurrency.lockutils [req-47a51e1f-8c27-4d36-a9c8-b1053c46f395 req-1563ee9b-6031-4e63-8d89-a431425f208e service nova] Acquiring lock "refresh_cache-3fa11db4-2f70-495e-a562-ac5f06153847" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1130.593538] env[63088]: DEBUG oslo_concurrency.lockutils [req-47a51e1f-8c27-4d36-a9c8-b1053c46f395 req-1563ee9b-6031-4e63-8d89-a431425f208e service nova] Acquired lock "refresh_cache-3fa11db4-2f70-495e-a562-ac5f06153847" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1130.593713] env[63088]: DEBUG nova.network.neutron [req-47a51e1f-8c27-4d36-a9c8-b1053c46f395 req-1563ee9b-6031-4e63-8d89-a431425f208e service nova] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Refreshing network info cache for port eb4d6233-2d9d-4756-8b9d-084bdfaeb3f7 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1130.728630] env[63088]: DEBUG nova.objects.base [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Object Instance<3fa11db4-2f70-495e-a562-ac5f06153847> lazy-loaded attributes: resources,numa_topology {{(pid=63088) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1130.769732] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5f07383-bb3a-4b5c-9a61-3028ba844716 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.777454] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5520d2ef-d59f-4028-b056-6068f7c4d686 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.806890] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4e76eea-2cad-411d-9f3c-490405c03938 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.815128] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c961733-925b-4247-b0e5-48bf2472d928 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.827833] env[63088]: DEBUG nova.compute.provider_tree [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1131.011834] env[63088]: DEBUG oslo_concurrency.lockutils [None req-91152bfa-3f14-496a-9e3c-24896547bba0 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Releasing lock "refresh_cache-2b87a9f1-5f10-43c2-8bc9-6d560ec88015" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1131.032669] env[63088]: INFO nova.compute.manager [-] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Took 1.38 seconds to deallocate network for instance. [ 1131.311340] env[63088]: DEBUG nova.network.neutron [req-47a51e1f-8c27-4d36-a9c8-b1053c46f395 req-1563ee9b-6031-4e63-8d89-a431425f208e service nova] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Updated VIF entry in instance network info cache for port eb4d6233-2d9d-4756-8b9d-084bdfaeb3f7. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1131.311662] env[63088]: DEBUG nova.network.neutron [req-47a51e1f-8c27-4d36-a9c8-b1053c46f395 req-1563ee9b-6031-4e63-8d89-a431425f208e service nova] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Updating instance_info_cache with network_info: [{"id": "eb4d6233-2d9d-4756-8b9d-084bdfaeb3f7", "address": "fa:16:3e:c2:db:1d", "network": {"id": "cc2c0304-8906-443a-a76e-e7a32158346f", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-1867758042-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.195", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b18ddddf2314d83addf550b8cb91977", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapeb4d6233-2d", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1131.330487] env[63088]: DEBUG nova.scheduler.client.report [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1131.516840] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-91152bfa-3f14-496a-9e3c-24896547bba0 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1131.516840] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3fb59e39-ab0c-4306-a190-a19f75dd7ccd {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.525504] env[63088]: DEBUG oslo_vmware.api [None req-91152bfa-3f14-496a-9e3c-24896547bba0 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1131.525504] env[63088]: value = "task-1285561" [ 1131.525504] env[63088]: _type = "Task" [ 1131.525504] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.534843] env[63088]: DEBUG oslo_vmware.api [None req-91152bfa-3f14-496a-9e3c-24896547bba0 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285561, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.538826] env[63088]: DEBUG oslo_concurrency.lockutils [None req-7424b985-8d50-4561-b4cf-cb870c1731d5 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1131.815550] env[63088]: DEBUG oslo_concurrency.lockutils [req-47a51e1f-8c27-4d36-a9c8-b1053c46f395 req-1563ee9b-6031-4e63-8d89-a431425f208e service nova] Releasing lock "refresh_cache-3fa11db4-2f70-495e-a562-ac5f06153847" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1131.815550] env[63088]: DEBUG nova.compute.manager [req-47a51e1f-8c27-4d36-a9c8-b1053c46f395 req-1563ee9b-6031-4e63-8d89-a431425f208e service nova] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Received event network-vif-deleted-f344e65c-4141-4c1d-a8e0-4aa1571a8e2f {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1131.836070] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.112s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1131.837745] env[63088]: DEBUG oslo_concurrency.lockutils [None req-7424b985-8d50-4561-b4cf-cb870c1731d5 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.299s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1131.837971] env[63088]: DEBUG nova.objects.instance [None req-7424b985-8d50-4561-b4cf-cb870c1731d5 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lazy-loading 'resources' on Instance uuid b25b5883-20c0-4cc9-8cc1-2c22a0f5be81 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1132.038030] env[63088]: DEBUG oslo_vmware.api [None req-91152bfa-3f14-496a-9e3c-24896547bba0 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285561, 'name': PowerOnVM_Task, 'duration_secs': 0.379076} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.038744] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-91152bfa-3f14-496a-9e3c-24896547bba0 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1132.038938] env[63088]: DEBUG nova.compute.manager [None req-91152bfa-3f14-496a-9e3c-24896547bba0 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1132.039692] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c3b088d-ed0e-4276-bc6d-e1ab661dff0f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.345723] env[63088]: DEBUG oslo_concurrency.lockutils [None req-d5ff234a-d6ff-455f-8893-ac0eb9577cf3 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Lock "3fa11db4-2f70-495e-a562-ac5f06153847" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 21.546s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1132.393501] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85c13153-3a94-4723-959e-6a50472234df {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.403030] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2031393d-533c-4a52-9b0d-0f7d61c7f342 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.431988] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0521f01-3523-4ed7-95b6-c00ff18d108f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.438948] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-664332c7-12b3-4618-b92d-1f18fe938ee5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.451940] env[63088]: DEBUG nova.compute.provider_tree [None req-7424b985-8d50-4561-b4cf-cb870c1731d5 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1132.529891] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Acquiring lock "3fa11db4-2f70-495e-a562-ac5f06153847" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1132.530153] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Lock "3fa11db4-2f70-495e-a562-ac5f06153847" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1132.530336] env[63088]: INFO nova.compute.manager [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Unshelving [ 1132.955965] env[63088]: DEBUG nova.scheduler.client.report [None req-7424b985-8d50-4561-b4cf-cb870c1731d5 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1133.326286] env[63088]: DEBUG oslo_concurrency.lockutils [None req-29800fc1-aa6e-40f8-a808-c5ac3bdcbee0 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquiring lock "2b87a9f1-5f10-43c2-8bc9-6d560ec88015" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1133.326557] env[63088]: DEBUG oslo_concurrency.lockutils [None req-29800fc1-aa6e-40f8-a808-c5ac3bdcbee0 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lock "2b87a9f1-5f10-43c2-8bc9-6d560ec88015" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1133.326763] env[63088]: DEBUG oslo_concurrency.lockutils [None req-29800fc1-aa6e-40f8-a808-c5ac3bdcbee0 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquiring lock "2b87a9f1-5f10-43c2-8bc9-6d560ec88015-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1133.326948] env[63088]: DEBUG oslo_concurrency.lockutils [None req-29800fc1-aa6e-40f8-a808-c5ac3bdcbee0 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lock "2b87a9f1-5f10-43c2-8bc9-6d560ec88015-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1133.327139] env[63088]: DEBUG oslo_concurrency.lockutils [None req-29800fc1-aa6e-40f8-a808-c5ac3bdcbee0 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lock "2b87a9f1-5f10-43c2-8bc9-6d560ec88015-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1133.329292] env[63088]: INFO nova.compute.manager [None req-29800fc1-aa6e-40f8-a808-c5ac3bdcbee0 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Terminating instance [ 1133.331009] env[63088]: DEBUG nova.compute.manager [None req-29800fc1-aa6e-40f8-a808-c5ac3bdcbee0 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1133.331212] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-29800fc1-aa6e-40f8-a808-c5ac3bdcbee0 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1133.332057] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5da0e8c-52a9-4aa0-b413-92e690aa8390 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.340859] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-29800fc1-aa6e-40f8-a808-c5ac3bdcbee0 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1133.341092] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c475a9b1-cc34-4d8e-9093-90431e5ba916 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.347251] env[63088]: DEBUG oslo_vmware.api [None req-29800fc1-aa6e-40f8-a808-c5ac3bdcbee0 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1133.347251] env[63088]: value = "task-1285562" [ 1133.347251] env[63088]: _type = "Task" [ 1133.347251] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.354509] env[63088]: DEBUG oslo_vmware.api [None req-29800fc1-aa6e-40f8-a808-c5ac3bdcbee0 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285562, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.460668] env[63088]: DEBUG oslo_concurrency.lockutils [None req-7424b985-8d50-4561-b4cf-cb870c1731d5 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.623s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1133.479030] env[63088]: INFO nova.scheduler.client.report [None req-7424b985-8d50-4561-b4cf-cb870c1731d5 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Deleted allocations for instance b25b5883-20c0-4cc9-8cc1-2c22a0f5be81 [ 1133.551072] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1133.551362] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1133.551573] env[63088]: DEBUG nova.objects.instance [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Lazy-loading 'pci_requests' on Instance uuid 3fa11db4-2f70-495e-a562-ac5f06153847 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1133.857704] env[63088]: DEBUG oslo_vmware.api [None req-29800fc1-aa6e-40f8-a808-c5ac3bdcbee0 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285562, 'name': PowerOffVM_Task, 'duration_secs': 0.179299} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.857980] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-29800fc1-aa6e-40f8-a808-c5ac3bdcbee0 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1133.858183] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-29800fc1-aa6e-40f8-a808-c5ac3bdcbee0 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Unregistering the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1133.858437] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-49897c9b-bf78-453d-aa59-bf5ce13dafa6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.913256] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-29800fc1-aa6e-40f8-a808-c5ac3bdcbee0 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Unregistered the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1133.913427] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-29800fc1-aa6e-40f8-a808-c5ac3bdcbee0 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Deleting contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1133.913618] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-29800fc1-aa6e-40f8-a808-c5ac3bdcbee0 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Deleting the datastore file [datastore1] 2b87a9f1-5f10-43c2-8bc9-6d560ec88015 {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1133.913908] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-94cdd8b3-8233-4689-b42f-b691ff01e844 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.920734] env[63088]: DEBUG oslo_vmware.api [None req-29800fc1-aa6e-40f8-a808-c5ac3bdcbee0 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1133.920734] env[63088]: value = "task-1285564" [ 1133.920734] env[63088]: _type = "Task" [ 1133.920734] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.928886] env[63088]: DEBUG oslo_vmware.api [None req-29800fc1-aa6e-40f8-a808-c5ac3bdcbee0 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285564, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.987163] env[63088]: DEBUG oslo_concurrency.lockutils [None req-7424b985-8d50-4561-b4cf-cb870c1731d5 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lock "b25b5883-20c0-4cc9-8cc1-2c22a0f5be81" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.453s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1134.055051] env[63088]: DEBUG nova.objects.instance [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Lazy-loading 'numa_topology' on Instance uuid 3fa11db4-2f70-495e-a562-ac5f06153847 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1134.430285] env[63088]: DEBUG oslo_vmware.api [None req-29800fc1-aa6e-40f8-a808-c5ac3bdcbee0 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285564, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.150016} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.430715] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-29800fc1-aa6e-40f8-a808-c5ac3bdcbee0 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1134.430715] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-29800fc1-aa6e-40f8-a808-c5ac3bdcbee0 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Deleted contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1134.430923] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-29800fc1-aa6e-40f8-a808-c5ac3bdcbee0 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1134.431124] env[63088]: INFO nova.compute.manager [None req-29800fc1-aa6e-40f8-a808-c5ac3bdcbee0 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1134.431371] env[63088]: DEBUG oslo.service.loopingcall [None req-29800fc1-aa6e-40f8-a808-c5ac3bdcbee0 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1134.431570] env[63088]: DEBUG nova.compute.manager [-] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1134.431660] env[63088]: DEBUG nova.network.neutron [-] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1134.557642] env[63088]: INFO nova.compute.claims [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1134.846466] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._sync_power_states {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1134.923626] env[63088]: DEBUG nova.compute.manager [req-d44dcff6-db65-40d5-bae8-98424b8cac4f req-aad6e7dc-68fc-455f-9600-ad9ce88c983c service nova] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Received event network-vif-deleted-f88388b7-bc70-4095-a792-8d155ef34534 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1134.923626] env[63088]: INFO nova.compute.manager [req-d44dcff6-db65-40d5-bae8-98424b8cac4f req-aad6e7dc-68fc-455f-9600-ad9ce88c983c service nova] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Neutron deleted interface f88388b7-bc70-4095-a792-8d155ef34534; detaching it from the instance and deleting it from the info cache [ 1134.923626] env[63088]: DEBUG nova.network.neutron [req-d44dcff6-db65-40d5-bae8-98424b8cac4f req-aad6e7dc-68fc-455f-9600-ad9ce88c983c service nova] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1135.350333] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Getting list of instances from cluster (obj){ [ 1135.350333] env[63088]: value = "domain-c8" [ 1135.350333] env[63088]: _type = "ClusterComputeResource" [ 1135.350333] env[63088]: } {{(pid=63088) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 1135.351445] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9527657c-5d6d-41b7-bdc4-d4e444a57036 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.360720] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Got total of 0 instances {{(pid=63088) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 1135.360898] env[63088]: WARNING nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] While synchronizing instance power states, found 2 instances in the database and 0 instances on the hypervisor. [ 1135.361048] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Triggering sync for uuid 2b87a9f1-5f10-43c2-8bc9-6d560ec88015 {{(pid=63088) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10336}} [ 1135.361245] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Triggering sync for uuid 3fa11db4-2f70-495e-a562-ac5f06153847 {{(pid=63088) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10336}} [ 1135.361555] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Acquiring lock "2b87a9f1-5f10-43c2-8bc9-6d560ec88015" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1135.361794] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Acquiring lock "3fa11db4-2f70-495e-a562-ac5f06153847" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1135.405404] env[63088]: DEBUG nova.network.neutron [-] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1135.427202] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-11c5f9f5-7564-4820-834f-c01b63b55ad5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.439698] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0225040-5d56-4841-afe6-0633b11f7a79 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.462339] env[63088]: DEBUG nova.compute.manager [req-d44dcff6-db65-40d5-bae8-98424b8cac4f req-aad6e7dc-68fc-455f-9600-ad9ce88c983c service nova] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Detach interface failed, port_id=f88388b7-bc70-4095-a792-8d155ef34534, reason: Instance 2b87a9f1-5f10-43c2-8bc9-6d560ec88015 could not be found. {{(pid=63088) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1135.518014] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquiring lock "7d8e3c6f-0328-405e-a869-d014caad8641" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1135.518357] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lock "7d8e3c6f-0328-405e-a869-d014caad8641" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1135.622444] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82886d17-54e9-4df6-9a36-c603f3e84d9d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.630189] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c2606bc-5475-49f7-805f-369e31cfd3d0 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.660282] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec62a131-7fe0-4062-a996-a40fde84e487 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.668122] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad023baf-f922-4379-974a-a0e946df254d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.681058] env[63088]: DEBUG nova.compute.provider_tree [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1135.908172] env[63088]: INFO nova.compute.manager [-] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Took 1.48 seconds to deallocate network for instance. [ 1136.021200] env[63088]: DEBUG nova.compute.manager [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1136.184753] env[63088]: DEBUG nova.scheduler.client.report [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1136.417354] env[63088]: DEBUG oslo_concurrency.lockutils [None req-29800fc1-aa6e-40f8-a808-c5ac3bdcbee0 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1136.543224] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1136.689934] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.138s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1136.692862] env[63088]: DEBUG oslo_concurrency.lockutils [None req-29800fc1-aa6e-40f8-a808-c5ac3bdcbee0 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.276s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1136.693261] env[63088]: DEBUG oslo_concurrency.lockutils [None req-29800fc1-aa6e-40f8-a808-c5ac3bdcbee0 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1136.695022] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.152s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1136.696598] env[63088]: INFO nova.compute.claims [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1136.719220] env[63088]: INFO nova.scheduler.client.report [None req-29800fc1-aa6e-40f8-a808-c5ac3bdcbee0 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Deleted allocations for instance 2b87a9f1-5f10-43c2-8bc9-6d560ec88015 [ 1136.745146] env[63088]: INFO nova.network.neutron [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Updating port eb4d6233-2d9d-4756-8b9d-084bdfaeb3f7 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1137.226203] env[63088]: DEBUG oslo_concurrency.lockutils [None req-29800fc1-aa6e-40f8-a808-c5ac3bdcbee0 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lock "2b87a9f1-5f10-43c2-8bc9-6d560ec88015" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.900s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1137.227110] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Lock "2b87a9f1-5f10-43c2-8bc9-6d560ec88015" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 1.866s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1137.227309] env[63088]: INFO nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] During sync_power_state the instance has a pending task (deleting). Skip. [ 1137.227483] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Lock "2b87a9f1-5f10-43c2-8bc9-6d560ec88015" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1137.751963] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afd54d6f-a865-4092-b8d1-c2a63f68c5c6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.760013] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a56a3244-a91d-4ffc-9fec-496a71a3c859 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.788867] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14c79f10-cbf5-4ad7-b3a1-8d3e81e2cb34 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.795717] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2d0dd59-f386-433c-b513-c1a3d9149114 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.809349] env[63088]: DEBUG nova.compute.provider_tree [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1138.184701] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquiring lock "f721d879-55a5-42d1-99c5-871d2e9103ec" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1138.185016] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lock "f721d879-55a5-42d1-99c5-871d2e9103ec" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1138.312398] env[63088]: DEBUG nova.scheduler.client.report [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1138.416321] env[63088]: DEBUG nova.compute.manager [req-4dd003e6-18e5-485e-8130-11f2e7d29f22 req-bbdf71eb-7c1b-4703-ba8a-0661713756f8 service nova] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Received event network-vif-plugged-eb4d6233-2d9d-4756-8b9d-084bdfaeb3f7 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1138.416321] env[63088]: DEBUG oslo_concurrency.lockutils [req-4dd003e6-18e5-485e-8130-11f2e7d29f22 req-bbdf71eb-7c1b-4703-ba8a-0661713756f8 service nova] Acquiring lock "3fa11db4-2f70-495e-a562-ac5f06153847-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1138.416321] env[63088]: DEBUG oslo_concurrency.lockutils [req-4dd003e6-18e5-485e-8130-11f2e7d29f22 req-bbdf71eb-7c1b-4703-ba8a-0661713756f8 service nova] Lock "3fa11db4-2f70-495e-a562-ac5f06153847-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1138.416321] env[63088]: DEBUG oslo_concurrency.lockutils [req-4dd003e6-18e5-485e-8130-11f2e7d29f22 req-bbdf71eb-7c1b-4703-ba8a-0661713756f8 service nova] Lock "3fa11db4-2f70-495e-a562-ac5f06153847-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1138.416321] env[63088]: DEBUG nova.compute.manager [req-4dd003e6-18e5-485e-8130-11f2e7d29f22 req-bbdf71eb-7c1b-4703-ba8a-0661713756f8 service nova] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] No waiting events found dispatching network-vif-plugged-eb4d6233-2d9d-4756-8b9d-084bdfaeb3f7 {{(pid=63088) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1138.416321] env[63088]: WARNING nova.compute.manager [req-4dd003e6-18e5-485e-8130-11f2e7d29f22 req-bbdf71eb-7c1b-4703-ba8a-0661713756f8 service nova] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Received unexpected event network-vif-plugged-eb4d6233-2d9d-4756-8b9d-084bdfaeb3f7 for instance with vm_state shelved_offloaded and task_state spawning. [ 1138.501374] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Acquiring lock "refresh_cache-3fa11db4-2f70-495e-a562-ac5f06153847" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1138.501374] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Acquired lock "refresh_cache-3fa11db4-2f70-495e-a562-ac5f06153847" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1138.501374] env[63088]: DEBUG nova.network.neutron [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1138.688366] env[63088]: DEBUG nova.compute.manager [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1138.817102] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.122s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1138.817938] env[63088]: DEBUG nova.compute.manager [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1139.206892] env[63088]: DEBUG nova.network.neutron [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Updating instance_info_cache with network_info: [{"id": "eb4d6233-2d9d-4756-8b9d-084bdfaeb3f7", "address": "fa:16:3e:c2:db:1d", "network": {"id": "cc2c0304-8906-443a-a76e-e7a32158346f", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1867758042-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.195", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b18ddddf2314d83addf550b8cb91977", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0f096917-a0cf-4add-a9d2-23ca1c723b3b", "external-id": "nsx-vlan-transportzone-894", "segmentation_id": 894, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeb4d6233-2d", "ovs_interfaceid": "eb4d6233-2d9d-4756-8b9d-084bdfaeb3f7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1139.210078] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1139.210327] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1139.211825] env[63088]: INFO nova.compute.claims [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1139.322045] env[63088]: DEBUG nova.compute.utils [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1139.323473] env[63088]: DEBUG nova.compute.manager [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1139.323734] env[63088]: DEBUG nova.network.neutron [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1139.369423] env[63088]: DEBUG nova.policy [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '787d7bd9f8e7490c9f24ff0e159bc98f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'dace8b5181b84623b08f903d12dfd31e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 1139.615345] env[63088]: DEBUG nova.network.neutron [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Successfully created port: 3e8c16d2-05c9-4554-8a2b-5cdb9b4b3a81 {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1139.709493] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Releasing lock "refresh_cache-3fa11db4-2f70-495e-a562-ac5f06153847" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1139.739078] env[63088]: DEBUG nova.virt.hardware [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='e436e7a9cc75a5babe98e1e7c27b3e2e',container_format='bare',created_at=2024-10-15T12:23:40Z,direct_url=,disk_format='vmdk',id=f1db6e74-0b10-490d-a958-6fa032fe3caf,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-1693433408-shelved',owner='2b18ddddf2314d83addf550b8cb91977',properties=ImageMetaProps,protected=,size=31668736,status='active',tags=,updated_at=2024-10-15T12:23:54Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1139.739330] env[63088]: DEBUG nova.virt.hardware [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1139.739494] env[63088]: DEBUG nova.virt.hardware [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1139.739680] env[63088]: DEBUG nova.virt.hardware [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1139.739832] env[63088]: DEBUG nova.virt.hardware [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1139.739984] env[63088]: DEBUG nova.virt.hardware [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1139.740322] env[63088]: DEBUG nova.virt.hardware [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1139.740495] env[63088]: DEBUG nova.virt.hardware [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1139.740669] env[63088]: DEBUG nova.virt.hardware [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1139.740838] env[63088]: DEBUG nova.virt.hardware [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1139.741054] env[63088]: DEBUG nova.virt.hardware [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1139.742255] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b833ee2b-1c10-44c8-b083-1d89116366f5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.749958] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48e2f595-9115-4e74-a953-cb27853aa430 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.764706] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c2:db:1d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0f096917-a0cf-4add-a9d2-23ca1c723b3b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'eb4d6233-2d9d-4756-8b9d-084bdfaeb3f7', 'vif_model': 'vmxnet3'}] {{(pid=63088) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1139.772792] env[63088]: DEBUG oslo.service.loopingcall [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1139.773329] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Creating VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1139.773549] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a77c8dbb-9b58-4641-817c-c87170fc1d95 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.792444] env[63088]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1139.792444] env[63088]: value = "task-1285565" [ 1139.792444] env[63088]: _type = "Task" [ 1139.792444] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1139.799442] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285565, 'name': CreateVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.826508] env[63088]: DEBUG nova.compute.manager [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1140.271882] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c57a4347-ee9c-4521-8f15-6387941d42c7 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.279966] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9ffc4b2-6006-4df3-8f10-e2e024775ef0 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.311828] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1aedf03-e8fa-4fec-a782-e905f7345263 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.319084] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285565, 'name': CreateVM_Task, 'duration_secs': 0.309129} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1140.320942] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Created VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1140.321736] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f1db6e74-0b10-490d-a958-6fa032fe3caf" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1140.321940] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f1db6e74-0b10-490d-a958-6fa032fe3caf" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1140.322364] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/f1db6e74-0b10-490d-a958-6fa032fe3caf" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1140.323658] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41f3813b-146d-40ec-9a2a-49153110b72f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.327295] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4103140e-4dde-4057-b553-30b1517f77f5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.344023] env[63088]: DEBUG nova.compute.provider_tree [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1140.345225] env[63088]: DEBUG oslo_vmware.api [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Waiting for the task: (returnval){ [ 1140.345225] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52c7fa82-7743-cfd9-2f6e-a04c52773512" [ 1140.345225] env[63088]: _type = "Task" [ 1140.345225] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1140.353934] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f1db6e74-0b10-490d-a958-6fa032fe3caf" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1140.354246] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Processing image f1db6e74-0b10-490d-a958-6fa032fe3caf {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1140.354511] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f1db6e74-0b10-490d-a958-6fa032fe3caf/f1db6e74-0b10-490d-a958-6fa032fe3caf.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1140.354659] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f1db6e74-0b10-490d-a958-6fa032fe3caf/f1db6e74-0b10-490d-a958-6fa032fe3caf.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1140.354839] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1140.355340] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b2449300-f50d-4ae2-b6a0-c23039cd7a79 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.372733] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1140.372931] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1140.373689] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-92ea67d2-f055-4b12-9166-82684f5738d0 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.378762] env[63088]: DEBUG oslo_vmware.api [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Waiting for the task: (returnval){ [ 1140.378762] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]529be0da-344e-3362-bb0d-e7233002020f" [ 1140.378762] env[63088]: _type = "Task" [ 1140.378762] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1140.386491] env[63088]: DEBUG oslo_vmware.api [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]529be0da-344e-3362-bb0d-e7233002020f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.438422] env[63088]: DEBUG nova.compute.manager [req-65f0fc9e-113f-4eaf-b893-401bd0485687 req-52563e53-6409-4940-9d36-0d02b545e42a service nova] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Received event network-changed-eb4d6233-2d9d-4756-8b9d-084bdfaeb3f7 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1140.438610] env[63088]: DEBUG nova.compute.manager [req-65f0fc9e-113f-4eaf-b893-401bd0485687 req-52563e53-6409-4940-9d36-0d02b545e42a service nova] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Refreshing instance network info cache due to event network-changed-eb4d6233-2d9d-4756-8b9d-084bdfaeb3f7. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1140.438823] env[63088]: DEBUG oslo_concurrency.lockutils [req-65f0fc9e-113f-4eaf-b893-401bd0485687 req-52563e53-6409-4940-9d36-0d02b545e42a service nova] Acquiring lock "refresh_cache-3fa11db4-2f70-495e-a562-ac5f06153847" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1140.438965] env[63088]: DEBUG oslo_concurrency.lockutils [req-65f0fc9e-113f-4eaf-b893-401bd0485687 req-52563e53-6409-4940-9d36-0d02b545e42a service nova] Acquired lock "refresh_cache-3fa11db4-2f70-495e-a562-ac5f06153847" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1140.439431] env[63088]: DEBUG nova.network.neutron [req-65f0fc9e-113f-4eaf-b893-401bd0485687 req-52563e53-6409-4940-9d36-0d02b545e42a service nova] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Refreshing network info cache for port eb4d6233-2d9d-4756-8b9d-084bdfaeb3f7 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1140.847391] env[63088]: DEBUG nova.scheduler.client.report [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1140.851211] env[63088]: DEBUG nova.compute.manager [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1140.876077] env[63088]: DEBUG nova.virt.hardware [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1140.876355] env[63088]: DEBUG nova.virt.hardware [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1140.876526] env[63088]: DEBUG nova.virt.hardware [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1140.876717] env[63088]: DEBUG nova.virt.hardware [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1140.876867] env[63088]: DEBUG nova.virt.hardware [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1140.877075] env[63088]: DEBUG nova.virt.hardware [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1140.877261] env[63088]: DEBUG nova.virt.hardware [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1140.877394] env[63088]: DEBUG nova.virt.hardware [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1140.877563] env[63088]: DEBUG nova.virt.hardware [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1140.877727] env[63088]: DEBUG nova.virt.hardware [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1140.877897] env[63088]: DEBUG nova.virt.hardware [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1140.879051] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4588d48-2b64-499a-8255-b442c9a39b9c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.892579] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be8fdb61-1e71-4f46-ae00-4fa1ed785b0d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.896858] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Preparing fetch location {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1140.897116] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Fetch image to [datastore1] OSTACK_IMG_4f787597-0f77-4644-b98a-ba15d59d6999/OSTACK_IMG_4f787597-0f77-4644-b98a-ba15d59d6999.vmdk {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1140.897335] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Downloading stream optimized image f1db6e74-0b10-490d-a958-6fa032fe3caf to [datastore1] OSTACK_IMG_4f787597-0f77-4644-b98a-ba15d59d6999/OSTACK_IMG_4f787597-0f77-4644-b98a-ba15d59d6999.vmdk on the data store datastore1 as vApp {{(pid=63088) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1140.897511] env[63088]: DEBUG nova.virt.vmwareapi.images [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Downloading image file data f1db6e74-0b10-490d-a958-6fa032fe3caf to the ESX as VM named 'OSTACK_IMG_4f787597-0f77-4644-b98a-ba15d59d6999' {{(pid=63088) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1140.974398] env[63088]: DEBUG oslo_vmware.rw_handles [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1140.974398] env[63088]: value = "resgroup-9" [ 1140.974398] env[63088]: _type = "ResourcePool" [ 1140.974398] env[63088]: }. {{(pid=63088) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1140.974779] env[63088]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-0e07f334-a084-4d17-8bc8-f36cb61d1a7d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.998965] env[63088]: DEBUG oslo_vmware.rw_handles [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Lease: (returnval){ [ 1140.998965] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52b2df79-7340-3de4-15db-1ebeeecfc348" [ 1140.998965] env[63088]: _type = "HttpNfcLease" [ 1140.998965] env[63088]: } obtained for vApp import into resource pool (val){ [ 1140.998965] env[63088]: value = "resgroup-9" [ 1140.998965] env[63088]: _type = "ResourcePool" [ 1140.998965] env[63088]: }. {{(pid=63088) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1140.999295] env[63088]: DEBUG oslo_vmware.api [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Waiting for the lease: (returnval){ [ 1140.999295] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52b2df79-7340-3de4-15db-1ebeeecfc348" [ 1140.999295] env[63088]: _type = "HttpNfcLease" [ 1140.999295] env[63088]: } to be ready. {{(pid=63088) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1141.006832] env[63088]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1141.006832] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52b2df79-7340-3de4-15db-1ebeeecfc348" [ 1141.006832] env[63088]: _type = "HttpNfcLease" [ 1141.006832] env[63088]: } is initializing. {{(pid=63088) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1141.180741] env[63088]: DEBUG nova.network.neutron [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Successfully updated port: 3e8c16d2-05c9-4554-8a2b-5cdb9b4b3a81 {{(pid=63088) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1141.247174] env[63088]: DEBUG nova.network.neutron [req-65f0fc9e-113f-4eaf-b893-401bd0485687 req-52563e53-6409-4940-9d36-0d02b545e42a service nova] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Updated VIF entry in instance network info cache for port eb4d6233-2d9d-4756-8b9d-084bdfaeb3f7. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1141.247794] env[63088]: DEBUG nova.network.neutron [req-65f0fc9e-113f-4eaf-b893-401bd0485687 req-52563e53-6409-4940-9d36-0d02b545e42a service nova] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Updating instance_info_cache with network_info: [{"id": "eb4d6233-2d9d-4756-8b9d-084bdfaeb3f7", "address": "fa:16:3e:c2:db:1d", "network": {"id": "cc2c0304-8906-443a-a76e-e7a32158346f", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1867758042-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.195", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b18ddddf2314d83addf550b8cb91977", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0f096917-a0cf-4add-a9d2-23ca1c723b3b", "external-id": "nsx-vlan-transportzone-894", "segmentation_id": 894, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeb4d6233-2d", "ovs_interfaceid": "eb4d6233-2d9d-4756-8b9d-084bdfaeb3f7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1141.354995] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.144s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1141.355621] env[63088]: DEBUG nova.compute.manager [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1141.509724] env[63088]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1141.509724] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52b2df79-7340-3de4-15db-1ebeeecfc348" [ 1141.509724] env[63088]: _type = "HttpNfcLease" [ 1141.509724] env[63088]: } is ready. {{(pid=63088) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1141.510061] env[63088]: DEBUG oslo_vmware.rw_handles [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1141.510061] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52b2df79-7340-3de4-15db-1ebeeecfc348" [ 1141.510061] env[63088]: _type = "HttpNfcLease" [ 1141.510061] env[63088]: }. {{(pid=63088) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1141.510785] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b12b5fa3-d594-4a45-ad6d-f5fcd8ecb326 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.518064] env[63088]: DEBUG oslo_vmware.rw_handles [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b93988-7e12-7732-a022-c8312d95745e/disk-0.vmdk from lease info. {{(pid=63088) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1141.518251] env[63088]: DEBUG oslo_vmware.rw_handles [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Creating HTTP connection to write to file with size = 31668736 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b93988-7e12-7732-a022-c8312d95745e/disk-0.vmdk. {{(pid=63088) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1141.581314] env[63088]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-11b4fcc1-0914-4bb3-9c31-d44dea6a7a4b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.683123] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquiring lock "refresh_cache-7d8e3c6f-0328-405e-a869-d014caad8641" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1141.683293] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquired lock "refresh_cache-7d8e3c6f-0328-405e-a869-d014caad8641" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1141.683480] env[63088]: DEBUG nova.network.neutron [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1141.750786] env[63088]: DEBUG oslo_concurrency.lockutils [req-65f0fc9e-113f-4eaf-b893-401bd0485687 req-52563e53-6409-4940-9d36-0d02b545e42a service nova] Releasing lock "refresh_cache-3fa11db4-2f70-495e-a562-ac5f06153847" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1141.861270] env[63088]: DEBUG nova.compute.utils [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1141.864264] env[63088]: DEBUG nova.compute.manager [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1141.864371] env[63088]: DEBUG nova.network.neutron [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1141.902408] env[63088]: DEBUG nova.policy [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6ffb8b92c6d649c4b39d8df1c1d0c36c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a7898ddafe0d41038a6ae4277f2c2f48', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 1142.171676] env[63088]: DEBUG nova.network.neutron [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Successfully created port: 154da9f6-d230-479a-88e0-f13ebc655493 {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1142.235432] env[63088]: DEBUG nova.network.neutron [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1142.367366] env[63088]: DEBUG nova.compute.manager [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1142.463195] env[63088]: DEBUG nova.network.neutron [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Updating instance_info_cache with network_info: [{"id": "3e8c16d2-05c9-4554-8a2b-5cdb9b4b3a81", "address": "fa:16:3e:af:74:ac", "network": {"id": "dfcbe153-bc01-4362-9247-de9f0b1c847b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-841757397-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dace8b5181b84623b08f903d12dfd31e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c68b7663-4f0e-47f0-ac7f-40c6d952f7bb", "external-id": "nsx-vlan-transportzone-696", "segmentation_id": 696, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3e8c16d2-05", "ovs_interfaceid": "3e8c16d2-05c9-4554-8a2b-5cdb9b4b3a81", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1142.467428] env[63088]: DEBUG nova.compute.manager [req-8c182c81-506f-472f-9750-850ffd0c6386 req-238a55c0-5d3b-404f-b18f-47c951295bfd service nova] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Received event network-vif-plugged-3e8c16d2-05c9-4554-8a2b-5cdb9b4b3a81 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1142.467428] env[63088]: DEBUG oslo_concurrency.lockutils [req-8c182c81-506f-472f-9750-850ffd0c6386 req-238a55c0-5d3b-404f-b18f-47c951295bfd service nova] Acquiring lock "7d8e3c6f-0328-405e-a869-d014caad8641-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1142.467428] env[63088]: DEBUG oslo_concurrency.lockutils [req-8c182c81-506f-472f-9750-850ffd0c6386 req-238a55c0-5d3b-404f-b18f-47c951295bfd service nova] Lock "7d8e3c6f-0328-405e-a869-d014caad8641-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1142.467428] env[63088]: DEBUG oslo_concurrency.lockutils [req-8c182c81-506f-472f-9750-850ffd0c6386 req-238a55c0-5d3b-404f-b18f-47c951295bfd service nova] Lock "7d8e3c6f-0328-405e-a869-d014caad8641-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1142.467554] env[63088]: DEBUG nova.compute.manager [req-8c182c81-506f-472f-9750-850ffd0c6386 req-238a55c0-5d3b-404f-b18f-47c951295bfd service nova] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] No waiting events found dispatching network-vif-plugged-3e8c16d2-05c9-4554-8a2b-5cdb9b4b3a81 {{(pid=63088) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1142.468199] env[63088]: WARNING nova.compute.manager [req-8c182c81-506f-472f-9750-850ffd0c6386 req-238a55c0-5d3b-404f-b18f-47c951295bfd service nova] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Received unexpected event network-vif-plugged-3e8c16d2-05c9-4554-8a2b-5cdb9b4b3a81 for instance with vm_state building and task_state spawning. [ 1142.468199] env[63088]: DEBUG nova.compute.manager [req-8c182c81-506f-472f-9750-850ffd0c6386 req-238a55c0-5d3b-404f-b18f-47c951295bfd service nova] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Received event network-changed-3e8c16d2-05c9-4554-8a2b-5cdb9b4b3a81 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1142.468199] env[63088]: DEBUG nova.compute.manager [req-8c182c81-506f-472f-9750-850ffd0c6386 req-238a55c0-5d3b-404f-b18f-47c951295bfd service nova] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Refreshing instance network info cache due to event network-changed-3e8c16d2-05c9-4554-8a2b-5cdb9b4b3a81. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1142.468361] env[63088]: DEBUG oslo_concurrency.lockutils [req-8c182c81-506f-472f-9750-850ffd0c6386 req-238a55c0-5d3b-404f-b18f-47c951295bfd service nova] Acquiring lock "refresh_cache-7d8e3c6f-0328-405e-a869-d014caad8641" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1142.649884] env[63088]: DEBUG oslo_vmware.rw_handles [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Completed reading data from the image iterator. {{(pid=63088) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1142.650187] env[63088]: DEBUG oslo_vmware.rw_handles [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b93988-7e12-7732-a022-c8312d95745e/disk-0.vmdk. {{(pid=63088) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1142.651213] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eeddf6ad-627c-4553-b558-9fcb015574bf {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.658286] env[63088]: DEBUG oslo_vmware.rw_handles [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b93988-7e12-7732-a022-c8312d95745e/disk-0.vmdk is in state: ready. {{(pid=63088) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1142.658464] env[63088]: DEBUG oslo_vmware.rw_handles [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b93988-7e12-7732-a022-c8312d95745e/disk-0.vmdk. {{(pid=63088) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1142.658709] env[63088]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-24aba5ed-75ab-44f6-a034-f2b6dffca47c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.854721] env[63088]: DEBUG oslo_vmware.rw_handles [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b93988-7e12-7732-a022-c8312d95745e/disk-0.vmdk. {{(pid=63088) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1142.854956] env[63088]: INFO nova.virt.vmwareapi.images [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Downloaded image file data f1db6e74-0b10-490d-a958-6fa032fe3caf [ 1142.855793] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebcc13d6-f7a2-449a-9bbe-f7f7e5cb1f2e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.870618] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4509374e-5689-449f-88fa-cefcab47d803 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.902087] env[63088]: INFO nova.virt.vmwareapi.images [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] The imported VM was unregistered [ 1142.904541] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Caching image {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1142.904774] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Creating directory with path [datastore1] devstack-image-cache_base/f1db6e74-0b10-490d-a958-6fa032fe3caf {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1142.905058] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1e9c1962-c53a-4663-99e0-6a841375f3c2 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.925508] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Created directory with path [datastore1] devstack-image-cache_base/f1db6e74-0b10-490d-a958-6fa032fe3caf {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1142.925699] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_4f787597-0f77-4644-b98a-ba15d59d6999/OSTACK_IMG_4f787597-0f77-4644-b98a-ba15d59d6999.vmdk to [datastore1] devstack-image-cache_base/f1db6e74-0b10-490d-a958-6fa032fe3caf/f1db6e74-0b10-490d-a958-6fa032fe3caf.vmdk. {{(pid=63088) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1142.925946] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-dc421d09-c9fe-40cb-ad4f-8a6894199f23 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.933207] env[63088]: DEBUG oslo_vmware.api [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Waiting for the task: (returnval){ [ 1142.933207] env[63088]: value = "task-1285568" [ 1142.933207] env[63088]: _type = "Task" [ 1142.933207] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1142.941269] env[63088]: DEBUG oslo_vmware.api [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285568, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1142.970198] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Releasing lock "refresh_cache-7d8e3c6f-0328-405e-a869-d014caad8641" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1142.970396] env[63088]: DEBUG nova.compute.manager [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Instance network_info: |[{"id": "3e8c16d2-05c9-4554-8a2b-5cdb9b4b3a81", "address": "fa:16:3e:af:74:ac", "network": {"id": "dfcbe153-bc01-4362-9247-de9f0b1c847b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-841757397-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dace8b5181b84623b08f903d12dfd31e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c68b7663-4f0e-47f0-ac7f-40c6d952f7bb", "external-id": "nsx-vlan-transportzone-696", "segmentation_id": 696, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3e8c16d2-05", "ovs_interfaceid": "3e8c16d2-05c9-4554-8a2b-5cdb9b4b3a81", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1142.970805] env[63088]: DEBUG oslo_concurrency.lockutils [req-8c182c81-506f-472f-9750-850ffd0c6386 req-238a55c0-5d3b-404f-b18f-47c951295bfd service nova] Acquired lock "refresh_cache-7d8e3c6f-0328-405e-a869-d014caad8641" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1142.971100] env[63088]: DEBUG nova.network.neutron [req-8c182c81-506f-472f-9750-850ffd0c6386 req-238a55c0-5d3b-404f-b18f-47c951295bfd service nova] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Refreshing network info cache for port 3e8c16d2-05c9-4554-8a2b-5cdb9b4b3a81 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1142.972542] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:af:74:ac', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c68b7663-4f0e-47f0-ac7f-40c6d952f7bb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3e8c16d2-05c9-4554-8a2b-5cdb9b4b3a81', 'vif_model': 'vmxnet3'}] {{(pid=63088) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1142.980728] env[63088]: DEBUG oslo.service.loopingcall [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1142.981898] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Creating VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1142.982217] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bc97c36a-3941-40c4-82c1-4ac1a489ab0e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.006268] env[63088]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1143.006268] env[63088]: value = "task-1285569" [ 1143.006268] env[63088]: _type = "Task" [ 1143.006268] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1143.014052] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285569, 'name': CreateVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1143.378346] env[63088]: DEBUG nova.compute.manager [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1143.406352] env[63088]: DEBUG nova.virt.hardware [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1143.406663] env[63088]: DEBUG nova.virt.hardware [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1143.406827] env[63088]: DEBUG nova.virt.hardware [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1143.407092] env[63088]: DEBUG nova.virt.hardware [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1143.407255] env[63088]: DEBUG nova.virt.hardware [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1143.407412] env[63088]: DEBUG nova.virt.hardware [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1143.407627] env[63088]: DEBUG nova.virt.hardware [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1143.407793] env[63088]: DEBUG nova.virt.hardware [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1143.407965] env[63088]: DEBUG nova.virt.hardware [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1143.408218] env[63088]: DEBUG nova.virt.hardware [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1143.408505] env[63088]: DEBUG nova.virt.hardware [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1143.409866] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-999d8648-f859-4094-ad36-4e3c15f3fde1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.420495] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc555e39-d434-48df-a09e-ed334406eda3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.443686] env[63088]: DEBUG oslo_vmware.api [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285568, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1143.518856] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285569, 'name': CreateVM_Task} progress is 99%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1143.861082] env[63088]: DEBUG nova.network.neutron [req-8c182c81-506f-472f-9750-850ffd0c6386 req-238a55c0-5d3b-404f-b18f-47c951295bfd service nova] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Updated VIF entry in instance network info cache for port 3e8c16d2-05c9-4554-8a2b-5cdb9b4b3a81. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1143.861082] env[63088]: DEBUG nova.network.neutron [req-8c182c81-506f-472f-9750-850ffd0c6386 req-238a55c0-5d3b-404f-b18f-47c951295bfd service nova] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Updating instance_info_cache with network_info: [{"id": "3e8c16d2-05c9-4554-8a2b-5cdb9b4b3a81", "address": "fa:16:3e:af:74:ac", "network": {"id": "dfcbe153-bc01-4362-9247-de9f0b1c847b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-841757397-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dace8b5181b84623b08f903d12dfd31e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c68b7663-4f0e-47f0-ac7f-40c6d952f7bb", "external-id": "nsx-vlan-transportzone-696", "segmentation_id": 696, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3e8c16d2-05", "ovs_interfaceid": "3e8c16d2-05c9-4554-8a2b-5cdb9b4b3a81", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1143.946037] env[63088]: DEBUG oslo_vmware.api [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285568, 'name': MoveVirtualDisk_Task} progress is 46%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1143.976531] env[63088]: DEBUG nova.network.neutron [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Successfully updated port: 154da9f6-d230-479a-88e0-f13ebc655493 {{(pid=63088) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1144.017057] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285569, 'name': CreateVM_Task, 'duration_secs': 0.575359} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1144.017209] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Created VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1144.017985] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1144.018174] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1144.018689] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1144.018882] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-768ea12c-130e-4fa4-92ab-85fe37182e9b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.024391] env[63088]: DEBUG oslo_vmware.api [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for the task: (returnval){ [ 1144.024391] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]526d3b59-2d6f-8bad-276a-7e0968e02f03" [ 1144.024391] env[63088]: _type = "Task" [ 1144.024391] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1144.032885] env[63088]: DEBUG oslo_vmware.api [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]526d3b59-2d6f-8bad-276a-7e0968e02f03, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.363881] env[63088]: DEBUG oslo_concurrency.lockutils [req-8c182c81-506f-472f-9750-850ffd0c6386 req-238a55c0-5d3b-404f-b18f-47c951295bfd service nova] Releasing lock "refresh_cache-7d8e3c6f-0328-405e-a869-d014caad8641" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1144.446201] env[63088]: DEBUG oslo_vmware.api [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285568, 'name': MoveVirtualDisk_Task} progress is 69%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.482095] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquiring lock "refresh_cache-f721d879-55a5-42d1-99c5-871d2e9103ec" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1144.482287] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquired lock "refresh_cache-f721d879-55a5-42d1-99c5-871d2e9103ec" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1144.482431] env[63088]: DEBUG nova.network.neutron [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1144.491323] env[63088]: DEBUG nova.compute.manager [req-abb73cdd-8a4a-4ada-9e44-59d627f6e204 req-43b84c1c-3e6e-4c14-9e2a-6915de659c32 service nova] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Received event network-vif-plugged-154da9f6-d230-479a-88e0-f13ebc655493 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1144.491516] env[63088]: DEBUG oslo_concurrency.lockutils [req-abb73cdd-8a4a-4ada-9e44-59d627f6e204 req-43b84c1c-3e6e-4c14-9e2a-6915de659c32 service nova] Acquiring lock "f721d879-55a5-42d1-99c5-871d2e9103ec-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1144.491731] env[63088]: DEBUG oslo_concurrency.lockutils [req-abb73cdd-8a4a-4ada-9e44-59d627f6e204 req-43b84c1c-3e6e-4c14-9e2a-6915de659c32 service nova] Lock "f721d879-55a5-42d1-99c5-871d2e9103ec-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1144.491943] env[63088]: DEBUG oslo_concurrency.lockutils [req-abb73cdd-8a4a-4ada-9e44-59d627f6e204 req-43b84c1c-3e6e-4c14-9e2a-6915de659c32 service nova] Lock "f721d879-55a5-42d1-99c5-871d2e9103ec-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1144.492091] env[63088]: DEBUG nova.compute.manager [req-abb73cdd-8a4a-4ada-9e44-59d627f6e204 req-43b84c1c-3e6e-4c14-9e2a-6915de659c32 service nova] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] No waiting events found dispatching network-vif-plugged-154da9f6-d230-479a-88e0-f13ebc655493 {{(pid=63088) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1144.492271] env[63088]: WARNING nova.compute.manager [req-abb73cdd-8a4a-4ada-9e44-59d627f6e204 req-43b84c1c-3e6e-4c14-9e2a-6915de659c32 service nova] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Received unexpected event network-vif-plugged-154da9f6-d230-479a-88e0-f13ebc655493 for instance with vm_state building and task_state spawning. [ 1144.492451] env[63088]: DEBUG nova.compute.manager [req-abb73cdd-8a4a-4ada-9e44-59d627f6e204 req-43b84c1c-3e6e-4c14-9e2a-6915de659c32 service nova] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Received event network-changed-154da9f6-d230-479a-88e0-f13ebc655493 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1144.492596] env[63088]: DEBUG nova.compute.manager [req-abb73cdd-8a4a-4ada-9e44-59d627f6e204 req-43b84c1c-3e6e-4c14-9e2a-6915de659c32 service nova] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Refreshing instance network info cache due to event network-changed-154da9f6-d230-479a-88e0-f13ebc655493. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1144.492764] env[63088]: DEBUG oslo_concurrency.lockutils [req-abb73cdd-8a4a-4ada-9e44-59d627f6e204 req-43b84c1c-3e6e-4c14-9e2a-6915de659c32 service nova] Acquiring lock "refresh_cache-f721d879-55a5-42d1-99c5-871d2e9103ec" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1144.536067] env[63088]: DEBUG oslo_vmware.api [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]526d3b59-2d6f-8bad-276a-7e0968e02f03, 'name': SearchDatastore_Task, 'duration_secs': 0.082697} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1144.536067] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1144.536361] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Processing image 1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1144.536645] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1144.536846] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1144.537149] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1144.537475] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-77a89612-f65d-48cb-a334-634c2b2c4742 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.552435] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1144.552699] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1144.553573] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ba197d14-5453-4eba-a00e-b518ba81fb1b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.559359] env[63088]: DEBUG oslo_vmware.api [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for the task: (returnval){ [ 1144.559359] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52ae988c-4087-b98f-bc14-2ded06bd4dc8" [ 1144.559359] env[63088]: _type = "Task" [ 1144.559359] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1144.568197] env[63088]: DEBUG oslo_vmware.api [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52ae988c-4087-b98f-bc14-2ded06bd4dc8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.946599] env[63088]: DEBUG oslo_vmware.api [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285568, 'name': MoveVirtualDisk_Task} progress is 91%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.017995] env[63088]: DEBUG nova.network.neutron [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1145.071570] env[63088]: DEBUG oslo_vmware.api [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52ae988c-4087-b98f-bc14-2ded06bd4dc8, 'name': SearchDatastore_Task, 'duration_secs': 0.077753} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1145.072483] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4c37f1c0-a529-4f38-80c4-f79323fb276c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.078204] env[63088]: DEBUG oslo_vmware.api [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for the task: (returnval){ [ 1145.078204] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52ed9007-2eb6-07dd-477d-33880dbefe29" [ 1145.078204] env[63088]: _type = "Task" [ 1145.078204] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1145.088475] env[63088]: DEBUG oslo_vmware.api [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52ed9007-2eb6-07dd-477d-33880dbefe29, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.165115] env[63088]: DEBUG nova.network.neutron [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Updating instance_info_cache with network_info: [{"id": "154da9f6-d230-479a-88e0-f13ebc655493", "address": "fa:16:3e:47:57:05", "network": {"id": "711d3d9a-86db-4aa1-87b5-481f1812dfa4", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-59376186-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a7898ddafe0d41038a6ae4277f2c2f48", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e41070eb-3ac1-4ca9-a3d0-fd65893a97de", "external-id": "nsx-vlan-transportzone-596", "segmentation_id": 596, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap154da9f6-d2", "ovs_interfaceid": "154da9f6-d230-479a-88e0-f13ebc655493", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1145.446753] env[63088]: DEBUG oslo_vmware.api [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285568, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.224611} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1145.447031] env[63088]: INFO nova.virt.vmwareapi.ds_util [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_4f787597-0f77-4644-b98a-ba15d59d6999/OSTACK_IMG_4f787597-0f77-4644-b98a-ba15d59d6999.vmdk to [datastore1] devstack-image-cache_base/f1db6e74-0b10-490d-a958-6fa032fe3caf/f1db6e74-0b10-490d-a958-6fa032fe3caf.vmdk. [ 1145.447229] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Cleaning up location [datastore1] OSTACK_IMG_4f787597-0f77-4644-b98a-ba15d59d6999 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1145.447394] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_4f787597-0f77-4644-b98a-ba15d59d6999 {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1145.447639] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-18c6bf9a-28d7-4ad0-b0fd-54f27033dfd5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.454279] env[63088]: DEBUG oslo_vmware.api [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Waiting for the task: (returnval){ [ 1145.454279] env[63088]: value = "task-1285570" [ 1145.454279] env[63088]: _type = "Task" [ 1145.454279] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1145.461367] env[63088]: DEBUG oslo_vmware.api [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285570, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.588803] env[63088]: DEBUG oslo_vmware.api [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52ed9007-2eb6-07dd-477d-33880dbefe29, 'name': SearchDatastore_Task, 'duration_secs': 0.016429} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1145.589147] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1145.589345] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] 7d8e3c6f-0328-405e-a869-d014caad8641/7d8e3c6f-0328-405e-a869-d014caad8641.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1145.589600] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bb699648-5ed0-42b2-96e6-eb481d535d70 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.595955] env[63088]: DEBUG oslo_vmware.api [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for the task: (returnval){ [ 1145.595955] env[63088]: value = "task-1285571" [ 1145.595955] env[63088]: _type = "Task" [ 1145.595955] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1145.603566] env[63088]: DEBUG oslo_vmware.api [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285571, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.668328] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Releasing lock "refresh_cache-f721d879-55a5-42d1-99c5-871d2e9103ec" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1145.668679] env[63088]: DEBUG nova.compute.manager [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Instance network_info: |[{"id": "154da9f6-d230-479a-88e0-f13ebc655493", "address": "fa:16:3e:47:57:05", "network": {"id": "711d3d9a-86db-4aa1-87b5-481f1812dfa4", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-59376186-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a7898ddafe0d41038a6ae4277f2c2f48", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e41070eb-3ac1-4ca9-a3d0-fd65893a97de", "external-id": "nsx-vlan-transportzone-596", "segmentation_id": 596, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap154da9f6-d2", "ovs_interfaceid": "154da9f6-d230-479a-88e0-f13ebc655493", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1145.669024] env[63088]: DEBUG oslo_concurrency.lockutils [req-abb73cdd-8a4a-4ada-9e44-59d627f6e204 req-43b84c1c-3e6e-4c14-9e2a-6915de659c32 service nova] Acquired lock "refresh_cache-f721d879-55a5-42d1-99c5-871d2e9103ec" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1145.669222] env[63088]: DEBUG nova.network.neutron [req-abb73cdd-8a4a-4ada-9e44-59d627f6e204 req-43b84c1c-3e6e-4c14-9e2a-6915de659c32 service nova] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Refreshing network info cache for port 154da9f6-d230-479a-88e0-f13ebc655493 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1145.670478] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:47:57:05', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e41070eb-3ac1-4ca9-a3d0-fd65893a97de', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '154da9f6-d230-479a-88e0-f13ebc655493', 'vif_model': 'vmxnet3'}] {{(pid=63088) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1145.677874] env[63088]: DEBUG oslo.service.loopingcall [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1145.678820] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Creating VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1145.679068] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fe4ab0cc-ed54-41b8-a105-541e26232362 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.698981] env[63088]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1145.698981] env[63088]: value = "task-1285572" [ 1145.698981] env[63088]: _type = "Task" [ 1145.698981] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1145.706558] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285572, 'name': CreateVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.968066] env[63088]: DEBUG oslo_vmware.api [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285570, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.035243} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1145.968618] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1145.968890] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f1db6e74-0b10-490d-a958-6fa032fe3caf/f1db6e74-0b10-490d-a958-6fa032fe3caf.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1145.969179] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/f1db6e74-0b10-490d-a958-6fa032fe3caf/f1db6e74-0b10-490d-a958-6fa032fe3caf.vmdk to [datastore1] 3fa11db4-2f70-495e-a562-ac5f06153847/3fa11db4-2f70-495e-a562-ac5f06153847.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1145.969469] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4d429003-9412-4a9d-b275-1c801fb97d22 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.976784] env[63088]: DEBUG oslo_vmware.api [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Waiting for the task: (returnval){ [ 1145.976784] env[63088]: value = "task-1285573" [ 1145.976784] env[63088]: _type = "Task" [ 1145.976784] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1145.985681] env[63088]: DEBUG oslo_vmware.api [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285573, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.105635] env[63088]: DEBUG oslo_vmware.api [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285571, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.482537} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1146.105861] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] 7d8e3c6f-0328-405e-a869-d014caad8641/7d8e3c6f-0328-405e-a869-d014caad8641.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1146.106094] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Extending root virtual disk to 1048576 {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1146.106352] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-844fbdf3-5cde-4f46-aacb-4ee6e7e16322 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.112347] env[63088]: DEBUG oslo_vmware.api [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for the task: (returnval){ [ 1146.112347] env[63088]: value = "task-1285574" [ 1146.112347] env[63088]: _type = "Task" [ 1146.112347] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1146.119217] env[63088]: DEBUG oslo_vmware.api [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285574, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.207876] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285572, 'name': CreateVM_Task, 'duration_secs': 0.396258} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1146.210044] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Created VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1146.211016] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1146.211207] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1146.211526] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1146.212073] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7d32f55b-229e-4ba2-98e2-d1abd140c97a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.216590] env[63088]: DEBUG oslo_vmware.api [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1146.216590] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52b36ac7-0aa1-3438-9868-1ec828564afb" [ 1146.216590] env[63088]: _type = "Task" [ 1146.216590] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1146.224603] env[63088]: DEBUG oslo_vmware.api [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52b36ac7-0aa1-3438-9868-1ec828564afb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.444348] env[63088]: DEBUG nova.network.neutron [req-abb73cdd-8a4a-4ada-9e44-59d627f6e204 req-43b84c1c-3e6e-4c14-9e2a-6915de659c32 service nova] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Updated VIF entry in instance network info cache for port 154da9f6-d230-479a-88e0-f13ebc655493. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1146.444758] env[63088]: DEBUG nova.network.neutron [req-abb73cdd-8a4a-4ada-9e44-59d627f6e204 req-43b84c1c-3e6e-4c14-9e2a-6915de659c32 service nova] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Updating instance_info_cache with network_info: [{"id": "154da9f6-d230-479a-88e0-f13ebc655493", "address": "fa:16:3e:47:57:05", "network": {"id": "711d3d9a-86db-4aa1-87b5-481f1812dfa4", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-59376186-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a7898ddafe0d41038a6ae4277f2c2f48", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e41070eb-3ac1-4ca9-a3d0-fd65893a97de", "external-id": "nsx-vlan-transportzone-596", "segmentation_id": 596, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap154da9f6-d2", "ovs_interfaceid": "154da9f6-d230-479a-88e0-f13ebc655493", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1146.488502] env[63088]: DEBUG oslo_vmware.api [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285573, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.622514] env[63088]: DEBUG oslo_vmware.api [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285574, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067771} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1146.622790] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Extended root virtual disk {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1146.623654] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea08fe56-3f4d-425c-a736-4eb6c66f94c4 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.646130] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Reconfiguring VM instance instance-00000069 to attach disk [datastore1] 7d8e3c6f-0328-405e-a869-d014caad8641/7d8e3c6f-0328-405e-a869-d014caad8641.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1146.646475] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-77462eec-dc3a-433a-a82b-9b5dcd63a518 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.666339] env[63088]: DEBUG oslo_vmware.api [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for the task: (returnval){ [ 1146.666339] env[63088]: value = "task-1285575" [ 1146.666339] env[63088]: _type = "Task" [ 1146.666339] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1146.674466] env[63088]: DEBUG oslo_vmware.api [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285575, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.726810] env[63088]: DEBUG oslo_vmware.api [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52b36ac7-0aa1-3438-9868-1ec828564afb, 'name': SearchDatastore_Task, 'duration_secs': 0.068732} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1146.727102] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1146.727350] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Processing image 1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1146.727599] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1146.727750] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1146.727934] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1146.728235] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-19471f51-a2e4-409f-bf69-452679d7339e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.743594] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1146.743920] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1146.744584] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c8a089b4-f930-4cba-81e2-4570e1e36aa1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.749973] env[63088]: DEBUG oslo_vmware.api [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1146.749973] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52e4398e-fc8d-7440-6cd1-0601a126f109" [ 1146.749973] env[63088]: _type = "Task" [ 1146.749973] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1146.757871] env[63088]: DEBUG oslo_vmware.api [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52e4398e-fc8d-7440-6cd1-0601a126f109, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.947896] env[63088]: DEBUG oslo_concurrency.lockutils [req-abb73cdd-8a4a-4ada-9e44-59d627f6e204 req-43b84c1c-3e6e-4c14-9e2a-6915de659c32 service nova] Releasing lock "refresh_cache-f721d879-55a5-42d1-99c5-871d2e9103ec" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1146.987732] env[63088]: DEBUG oslo_vmware.api [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285573, 'name': CopyVirtualDisk_Task} progress is 43%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.177020] env[63088]: DEBUG oslo_vmware.api [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285575, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.260380] env[63088]: DEBUG oslo_vmware.api [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52e4398e-fc8d-7440-6cd1-0601a126f109, 'name': SearchDatastore_Task, 'duration_secs': 0.076107} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1147.261262] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e6552c8f-99fa-46b9-88e4-b18f1bd6b9cc {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.266811] env[63088]: DEBUG oslo_vmware.api [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1147.266811] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52c3edde-16d2-0a63-60f3-102a7bf6bbee" [ 1147.266811] env[63088]: _type = "Task" [ 1147.266811] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1147.275223] env[63088]: DEBUG oslo_vmware.api [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52c3edde-16d2-0a63-60f3-102a7bf6bbee, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.489324] env[63088]: DEBUG oslo_vmware.api [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285573, 'name': CopyVirtualDisk_Task} progress is 66%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.677748] env[63088]: DEBUG oslo_vmware.api [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285575, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.777554] env[63088]: DEBUG oslo_vmware.api [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52c3edde-16d2-0a63-60f3-102a7bf6bbee, 'name': SearchDatastore_Task, 'duration_secs': 0.076577} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1147.777837] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1147.778143] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] f721d879-55a5-42d1-99c5-871d2e9103ec/f721d879-55a5-42d1-99c5-871d2e9103ec.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1147.778430] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2876b04c-b926-43ea-bf6b-3ad6a90285e1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.785636] env[63088]: DEBUG oslo_vmware.api [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1147.785636] env[63088]: value = "task-1285576" [ 1147.785636] env[63088]: _type = "Task" [ 1147.785636] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1147.793518] env[63088]: DEBUG oslo_vmware.api [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285576, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.990578] env[63088]: DEBUG oslo_vmware.api [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285573, 'name': CopyVirtualDisk_Task} progress is 88%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.180383] env[63088]: DEBUG oslo_vmware.api [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285575, 'name': ReconfigVM_Task, 'duration_secs': 1.215367} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1148.180685] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Reconfigured VM instance instance-00000069 to attach disk [datastore1] 7d8e3c6f-0328-405e-a869-d014caad8641/7d8e3c6f-0328-405e-a869-d014caad8641.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1148.181423] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-244ea023-3971-438b-8cc0-24cbf6066c65 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.188324] env[63088]: DEBUG oslo_vmware.api [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for the task: (returnval){ [ 1148.188324] env[63088]: value = "task-1285577" [ 1148.188324] env[63088]: _type = "Task" [ 1148.188324] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1148.197301] env[63088]: DEBUG oslo_vmware.api [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285577, 'name': Rename_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.295636] env[63088]: DEBUG oslo_vmware.api [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285576, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.490884] env[63088]: DEBUG oslo_vmware.api [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285573, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.698142] env[63088]: DEBUG oslo_vmware.api [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285577, 'name': Rename_Task, 'duration_secs': 0.137404} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1148.698555] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1148.699719] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-39cef6c4-c550-4d45-9fe5-caa933b2e1a6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.704707] env[63088]: DEBUG oslo_vmware.api [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for the task: (returnval){ [ 1148.704707] env[63088]: value = "task-1285578" [ 1148.704707] env[63088]: _type = "Task" [ 1148.704707] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1148.711673] env[63088]: DEBUG oslo_vmware.api [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285578, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.795466] env[63088]: DEBUG oslo_vmware.api [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285576, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.991970] env[63088]: DEBUG oslo_vmware.api [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285573, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.214068] env[63088]: DEBUG oslo_vmware.api [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285578, 'name': PowerOnVM_Task, 'duration_secs': 0.472713} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1149.214379] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1149.214601] env[63088]: INFO nova.compute.manager [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Took 8.36 seconds to spawn the instance on the hypervisor. [ 1149.214787] env[63088]: DEBUG nova.compute.manager [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1149.215552] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f757a76-1d93-4c93-b9d0-924eaf766a70 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.297051] env[63088]: DEBUG oslo_vmware.api [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285576, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.493116] env[63088]: DEBUG oslo_vmware.api [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285573, 'name': CopyVirtualDisk_Task, 'duration_secs': 3.252194} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1149.493426] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/f1db6e74-0b10-490d-a958-6fa032fe3caf/f1db6e74-0b10-490d-a958-6fa032fe3caf.vmdk to [datastore1] 3fa11db4-2f70-495e-a562-ac5f06153847/3fa11db4-2f70-495e-a562-ac5f06153847.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1149.494272] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64deb4d0-c886-4cc6-a265-e12f507e5b49 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.517165] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Reconfiguring VM instance instance-00000068 to attach disk [datastore1] 3fa11db4-2f70-495e-a562-ac5f06153847/3fa11db4-2f70-495e-a562-ac5f06153847.vmdk or device None with type streamOptimized {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1149.517899] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a95d61e1-1eea-402f-bf4f-aa2273d7a5c9 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.537937] env[63088]: DEBUG oslo_vmware.api [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Waiting for the task: (returnval){ [ 1149.537937] env[63088]: value = "task-1285579" [ 1149.537937] env[63088]: _type = "Task" [ 1149.537937] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1149.546485] env[63088]: DEBUG oslo_vmware.api [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285579, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.738168] env[63088]: INFO nova.compute.manager [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Took 13.21 seconds to build instance. [ 1149.798952] env[63088]: DEBUG oslo_vmware.api [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285576, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.8851} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1149.799203] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] f721d879-55a5-42d1-99c5-871d2e9103ec/f721d879-55a5-42d1-99c5-871d2e9103ec.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1149.799424] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Extending root virtual disk to 1048576 {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1149.799680] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e8f5db8c-3e9c-4bfb-a167-0e0fda9e99f2 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.806198] env[63088]: DEBUG oslo_vmware.api [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1149.806198] env[63088]: value = "task-1285580" [ 1149.806198] env[63088]: _type = "Task" [ 1149.806198] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1149.813460] env[63088]: DEBUG oslo_vmware.api [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285580, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.008781] env[63088]: DEBUG nova.compute.manager [req-bce4eb54-d7e7-4941-9bb8-2301d23382c7 req-dd16e0ba-9b7e-4697-8ac1-96649412e258 service nova] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Received event network-changed-3e8c16d2-05c9-4554-8a2b-5cdb9b4b3a81 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1150.009041] env[63088]: DEBUG nova.compute.manager [req-bce4eb54-d7e7-4941-9bb8-2301d23382c7 req-dd16e0ba-9b7e-4697-8ac1-96649412e258 service nova] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Refreshing instance network info cache due to event network-changed-3e8c16d2-05c9-4554-8a2b-5cdb9b4b3a81. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1150.009226] env[63088]: DEBUG oslo_concurrency.lockutils [req-bce4eb54-d7e7-4941-9bb8-2301d23382c7 req-dd16e0ba-9b7e-4697-8ac1-96649412e258 service nova] Acquiring lock "refresh_cache-7d8e3c6f-0328-405e-a869-d014caad8641" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1150.009374] env[63088]: DEBUG oslo_concurrency.lockutils [req-bce4eb54-d7e7-4941-9bb8-2301d23382c7 req-dd16e0ba-9b7e-4697-8ac1-96649412e258 service nova] Acquired lock "refresh_cache-7d8e3c6f-0328-405e-a869-d014caad8641" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1150.009542] env[63088]: DEBUG nova.network.neutron [req-bce4eb54-d7e7-4941-9bb8-2301d23382c7 req-dd16e0ba-9b7e-4697-8ac1-96649412e258 service nova] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Refreshing network info cache for port 3e8c16d2-05c9-4554-8a2b-5cdb9b4b3a81 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1150.047149] env[63088]: DEBUG oslo_vmware.api [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285579, 'name': ReconfigVM_Task, 'duration_secs': 0.344253} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1150.047423] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Reconfigured VM instance instance-00000068 to attach disk [datastore1] 3fa11db4-2f70-495e-a562-ac5f06153847/3fa11db4-2f70-495e-a562-ac5f06153847.vmdk or device None with type streamOptimized {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1150.048035] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d46d75e2-1859-46e3-94e3-49203138d02f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.055417] env[63088]: DEBUG oslo_vmware.api [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Waiting for the task: (returnval){ [ 1150.055417] env[63088]: value = "task-1285581" [ 1150.055417] env[63088]: _type = "Task" [ 1150.055417] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1150.063020] env[63088]: DEBUG oslo_vmware.api [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285581, 'name': Rename_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.240973] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a74e345a-2c38-4c6f-b83d-88cae66aa29b tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lock "7d8e3c6f-0328-405e-a869-d014caad8641" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.722s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1150.316103] env[63088]: DEBUG oslo_vmware.api [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285580, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.097015} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1150.316246] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Extended root virtual disk {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1150.316975] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-605678b1-aaac-45ec-8ac2-f23a35180d6f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.337846] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Reconfiguring VM instance instance-0000006a to attach disk [datastore1] f721d879-55a5-42d1-99c5-871d2e9103ec/f721d879-55a5-42d1-99c5-871d2e9103ec.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1150.338085] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-420f9f08-fee0-4230-a553-682238f2d926 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.356016] env[63088]: DEBUG oslo_vmware.api [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1150.356016] env[63088]: value = "task-1285582" [ 1150.356016] env[63088]: _type = "Task" [ 1150.356016] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1150.362899] env[63088]: DEBUG oslo_vmware.api [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285582, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.565533] env[63088]: DEBUG oslo_vmware.api [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285581, 'name': Rename_Task, 'duration_secs': 0.159647} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1150.565868] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1150.566222] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-488eaf77-d7c9-4601-97ee-8f4df0f5ffb1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.573799] env[63088]: DEBUG oslo_vmware.api [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Waiting for the task: (returnval){ [ 1150.573799] env[63088]: value = "task-1285583" [ 1150.573799] env[63088]: _type = "Task" [ 1150.573799] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1150.581880] env[63088]: DEBUG oslo_vmware.api [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285583, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.777993] env[63088]: DEBUG nova.network.neutron [req-bce4eb54-d7e7-4941-9bb8-2301d23382c7 req-dd16e0ba-9b7e-4697-8ac1-96649412e258 service nova] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Updated VIF entry in instance network info cache for port 3e8c16d2-05c9-4554-8a2b-5cdb9b4b3a81. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1150.778409] env[63088]: DEBUG nova.network.neutron [req-bce4eb54-d7e7-4941-9bb8-2301d23382c7 req-dd16e0ba-9b7e-4697-8ac1-96649412e258 service nova] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Updating instance_info_cache with network_info: [{"id": "3e8c16d2-05c9-4554-8a2b-5cdb9b4b3a81", "address": "fa:16:3e:af:74:ac", "network": {"id": "dfcbe153-bc01-4362-9247-de9f0b1c847b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-841757397-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.152", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dace8b5181b84623b08f903d12dfd31e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c68b7663-4f0e-47f0-ac7f-40c6d952f7bb", "external-id": "nsx-vlan-transportzone-696", "segmentation_id": 696, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3e8c16d2-05", "ovs_interfaceid": "3e8c16d2-05c9-4554-8a2b-5cdb9b4b3a81", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1150.865682] env[63088]: DEBUG oslo_vmware.api [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285582, 'name': ReconfigVM_Task, 'duration_secs': 0.25963} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1150.865955] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Reconfigured VM instance instance-0000006a to attach disk [datastore1] f721d879-55a5-42d1-99c5-871d2e9103ec/f721d879-55a5-42d1-99c5-871d2e9103ec.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1150.866602] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9880fc78-a8bb-4aa0-b508-7206f8d39cbf {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.872670] env[63088]: DEBUG oslo_vmware.api [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1150.872670] env[63088]: value = "task-1285584" [ 1150.872670] env[63088]: _type = "Task" [ 1150.872670] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1150.882244] env[63088]: DEBUG oslo_vmware.api [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285584, 'name': Rename_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.085251] env[63088]: DEBUG oslo_vmware.api [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285583, 'name': PowerOnVM_Task, 'duration_secs': 0.443449} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1151.085627] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1151.182715] env[63088]: DEBUG nova.compute.manager [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1151.183832] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb6cf5a7-4b21-42fb-af97-06eec98e5563 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.281264] env[63088]: DEBUG oslo_concurrency.lockutils [req-bce4eb54-d7e7-4941-9bb8-2301d23382c7 req-dd16e0ba-9b7e-4697-8ac1-96649412e258 service nova] Releasing lock "refresh_cache-7d8e3c6f-0328-405e-a869-d014caad8641" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1151.382732] env[63088]: DEBUG oslo_vmware.api [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285584, 'name': Rename_Task, 'duration_secs': 0.192132} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1151.382935] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1151.383235] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9c9dc926-ca8f-405b-86ad-906207c343f7 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.389551] env[63088]: DEBUG oslo_vmware.api [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1151.389551] env[63088]: value = "task-1285585" [ 1151.389551] env[63088]: _type = "Task" [ 1151.389551] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1151.396796] env[63088]: DEBUG oslo_vmware.api [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285585, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.703093] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0892c6cb-9ae9-45a6-8dbf-8069656cb3e2 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Lock "3fa11db4-2f70-495e-a562-ac5f06153847" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 19.172s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1151.703609] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Lock "3fa11db4-2f70-495e-a562-ac5f06153847" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 16.342s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1151.703809] env[63088]: INFO nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] During sync_power_state the instance has a pending task (spawning). Skip. [ 1151.704073] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Lock "3fa11db4-2f70-495e-a562-ac5f06153847" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1151.902100] env[63088]: DEBUG oslo_vmware.api [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285585, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.401051] env[63088]: DEBUG oslo_vmware.api [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285585, 'name': PowerOnVM_Task, 'duration_secs': 0.569102} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1152.401441] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1152.401562] env[63088]: INFO nova.compute.manager [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Took 9.02 seconds to spawn the instance on the hypervisor. [ 1152.401715] env[63088]: DEBUG nova.compute.manager [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1152.402477] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d814face-efb3-4feb-9365-1d591a8ecd55 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.919344] env[63088]: INFO nova.compute.manager [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Took 13.72 seconds to build instance. [ 1153.420706] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6e271f59-a8c9-4c49-bb52-7389caf0e18d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lock "f721d879-55a5-42d1-99c5-871d2e9103ec" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.235s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1153.549665] env[63088]: DEBUG nova.compute.manager [req-3c595d1a-2df0-4988-b885-2b78400cbac6 req-0bf43a8f-b36a-4e00-b0dd-0d7e33d0a1f5 service nova] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Received event network-changed-154da9f6-d230-479a-88e0-f13ebc655493 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1153.549899] env[63088]: DEBUG nova.compute.manager [req-3c595d1a-2df0-4988-b885-2b78400cbac6 req-0bf43a8f-b36a-4e00-b0dd-0d7e33d0a1f5 service nova] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Refreshing instance network info cache due to event network-changed-154da9f6-d230-479a-88e0-f13ebc655493. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1153.550122] env[63088]: DEBUG oslo_concurrency.lockutils [req-3c595d1a-2df0-4988-b885-2b78400cbac6 req-0bf43a8f-b36a-4e00-b0dd-0d7e33d0a1f5 service nova] Acquiring lock "refresh_cache-f721d879-55a5-42d1-99c5-871d2e9103ec" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1153.550271] env[63088]: DEBUG oslo_concurrency.lockutils [req-3c595d1a-2df0-4988-b885-2b78400cbac6 req-0bf43a8f-b36a-4e00-b0dd-0d7e33d0a1f5 service nova] Acquired lock "refresh_cache-f721d879-55a5-42d1-99c5-871d2e9103ec" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1153.550439] env[63088]: DEBUG nova.network.neutron [req-3c595d1a-2df0-4988-b885-2b78400cbac6 req-0bf43a8f-b36a-4e00-b0dd-0d7e33d0a1f5 service nova] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Refreshing network info cache for port 154da9f6-d230-479a-88e0-f13ebc655493 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1154.259692] env[63088]: DEBUG nova.network.neutron [req-3c595d1a-2df0-4988-b885-2b78400cbac6 req-0bf43a8f-b36a-4e00-b0dd-0d7e33d0a1f5 service nova] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Updated VIF entry in instance network info cache for port 154da9f6-d230-479a-88e0-f13ebc655493. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1154.260085] env[63088]: DEBUG nova.network.neutron [req-3c595d1a-2df0-4988-b885-2b78400cbac6 req-0bf43a8f-b36a-4e00-b0dd-0d7e33d0a1f5 service nova] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Updating instance_info_cache with network_info: [{"id": "154da9f6-d230-479a-88e0-f13ebc655493", "address": "fa:16:3e:47:57:05", "network": {"id": "711d3d9a-86db-4aa1-87b5-481f1812dfa4", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-59376186-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.225", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a7898ddafe0d41038a6ae4277f2c2f48", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e41070eb-3ac1-4ca9-a3d0-fd65893a97de", "external-id": "nsx-vlan-transportzone-596", "segmentation_id": 596, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap154da9f6-d2", "ovs_interfaceid": "154da9f6-d230-479a-88e0-f13ebc655493", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1154.762765] env[63088]: DEBUG oslo_concurrency.lockutils [req-3c595d1a-2df0-4988-b885-2b78400cbac6 req-0bf43a8f-b36a-4e00-b0dd-0d7e33d0a1f5 service nova] Releasing lock "refresh_cache-f721d879-55a5-42d1-99c5-871d2e9103ec" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1155.120486] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1156.604676] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1156.605113] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1156.605113] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63088) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1157.604842] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1161.600131] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1161.603802] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1161.603957] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Starting heal instance info cache {{(pid=63088) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1162.610060] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Didn't find any instances for network info cache update. {{(pid=63088) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10012}} [ 1162.610461] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1162.610461] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1162.610628] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager.update_available_resource {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1163.113902] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1163.114189] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1163.114373] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1163.114524] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63088) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1163.115472] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d85fa4a6-612a-4199-8822-02b9cead4a74 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.123989] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb262ad8-bb62-485e-a77f-cc9b3d2a98ab {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.138181] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e02c8420-5224-4ac0-8c01-83ef177ab14f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.144439] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f9b5710-48d6-4fc6-8ecf-b267d752b779 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.172787] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180856MB free_disk=140GB free_vcpus=48 pci_devices=None {{(pid=63088) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1163.172938] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1163.173145] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1164.198051] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance 3fa11db4-2f70-495e-a562-ac5f06153847 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1164.198276] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance 7d8e3c6f-0328-405e-a869-d014caad8641 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1164.198346] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance f721d879-55a5-42d1-99c5-871d2e9103ec actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1164.198899] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Total usable vcpus: 48, total allocated vcpus: 3 {{(pid=63088) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1164.198899] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1088MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=3 pci_stats=[] {{(pid=63088) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1164.243215] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c1b8fed-9541-4fbb-8000-b86a548b6e26 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.250391] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17e7869d-e11c-4d18-a7c2-ba63b552787d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.280838] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0634c7bf-d924-412b-9b8a-0e078154bc22 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.287850] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce5cba3b-3607-4ca9-9e7f-76b9db45b941 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.300416] env[63088]: DEBUG nova.compute.provider_tree [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1164.803981] env[63088]: DEBUG nova.scheduler.client.report [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1165.309065] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63088) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1165.309513] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.136s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1187.339595] env[63088]: DEBUG oslo_concurrency.lockutils [None req-84501e08-4dc3-4bd2-8840-2c26a498d329 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquiring lock "7d8e3c6f-0328-405e-a869-d014caad8641" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1187.339996] env[63088]: DEBUG oslo_concurrency.lockutils [None req-84501e08-4dc3-4bd2-8840-2c26a498d329 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lock "7d8e3c6f-0328-405e-a869-d014caad8641" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1187.340072] env[63088]: DEBUG nova.compute.manager [None req-84501e08-4dc3-4bd2-8840-2c26a498d329 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1187.340937] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da9a674f-2b05-4914-ae7f-b1456a9e297c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.347882] env[63088]: DEBUG nova.compute.manager [None req-84501e08-4dc3-4bd2-8840-2c26a498d329 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=63088) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1187.348450] env[63088]: DEBUG nova.objects.instance [None req-84501e08-4dc3-4bd2-8840-2c26a498d329 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lazy-loading 'flavor' on Instance uuid 7d8e3c6f-0328-405e-a869-d014caad8641 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1187.854805] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-84501e08-4dc3-4bd2-8840-2c26a498d329 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1187.855113] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-037925f7-9d05-49b5-a91e-b408a3846483 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.862938] env[63088]: DEBUG oslo_vmware.api [None req-84501e08-4dc3-4bd2-8840-2c26a498d329 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for the task: (returnval){ [ 1187.862938] env[63088]: value = "task-1285586" [ 1187.862938] env[63088]: _type = "Task" [ 1187.862938] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1187.872254] env[63088]: DEBUG oslo_vmware.api [None req-84501e08-4dc3-4bd2-8840-2c26a498d329 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285586, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1188.373052] env[63088]: DEBUG oslo_vmware.api [None req-84501e08-4dc3-4bd2-8840-2c26a498d329 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285586, 'name': PowerOffVM_Task, 'duration_secs': 0.169593} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1188.373052] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-84501e08-4dc3-4bd2-8840-2c26a498d329 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1188.373052] env[63088]: DEBUG nova.compute.manager [None req-84501e08-4dc3-4bd2-8840-2c26a498d329 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1188.373537] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06f34429-628a-4beb-9a70-2ed4c11a900f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.401800] env[63088]: DEBUG oslo_concurrency.lockutils [None req-9495a91e-d90a-4c5b-ba81-2200f8356cb8 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Acquiring lock "3fa11db4-2f70-495e-a562-ac5f06153847" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1188.402070] env[63088]: DEBUG oslo_concurrency.lockutils [None req-9495a91e-d90a-4c5b-ba81-2200f8356cb8 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Lock "3fa11db4-2f70-495e-a562-ac5f06153847" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1188.402289] env[63088]: DEBUG oslo_concurrency.lockutils [None req-9495a91e-d90a-4c5b-ba81-2200f8356cb8 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Acquiring lock "3fa11db4-2f70-495e-a562-ac5f06153847-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1188.402479] env[63088]: DEBUG oslo_concurrency.lockutils [None req-9495a91e-d90a-4c5b-ba81-2200f8356cb8 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Lock "3fa11db4-2f70-495e-a562-ac5f06153847-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1188.402653] env[63088]: DEBUG oslo_concurrency.lockutils [None req-9495a91e-d90a-4c5b-ba81-2200f8356cb8 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Lock "3fa11db4-2f70-495e-a562-ac5f06153847-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1188.404815] env[63088]: INFO nova.compute.manager [None req-9495a91e-d90a-4c5b-ba81-2200f8356cb8 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Terminating instance [ 1188.406329] env[63088]: DEBUG nova.compute.manager [None req-9495a91e-d90a-4c5b-ba81-2200f8356cb8 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1188.406525] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-9495a91e-d90a-4c5b-ba81-2200f8356cb8 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1188.407281] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2bad711-c294-4f4a-819f-1750fcf3afe3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.414998] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-9495a91e-d90a-4c5b-ba81-2200f8356cb8 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1188.415235] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-891574ce-310f-4b9a-9c8a-ab267f1b23b5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.421330] env[63088]: DEBUG oslo_vmware.api [None req-9495a91e-d90a-4c5b-ba81-2200f8356cb8 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Waiting for the task: (returnval){ [ 1188.421330] env[63088]: value = "task-1285587" [ 1188.421330] env[63088]: _type = "Task" [ 1188.421330] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1188.428494] env[63088]: DEBUG oslo_vmware.api [None req-9495a91e-d90a-4c5b-ba81-2200f8356cb8 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285587, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1188.884456] env[63088]: DEBUG oslo_concurrency.lockutils [None req-84501e08-4dc3-4bd2-8840-2c26a498d329 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lock "7d8e3c6f-0328-405e-a869-d014caad8641" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.544s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1188.931875] env[63088]: DEBUG oslo_vmware.api [None req-9495a91e-d90a-4c5b-ba81-2200f8356cb8 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285587, 'name': PowerOffVM_Task, 'duration_secs': 0.194918} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1188.932308] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-9495a91e-d90a-4c5b-ba81-2200f8356cb8 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1188.932572] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-9495a91e-d90a-4c5b-ba81-2200f8356cb8 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Unregistering the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1188.932882] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f7f734ae-077b-4d16-9045-2fa1b80922fd {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.990348] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-9495a91e-d90a-4c5b-ba81-2200f8356cb8 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Unregistered the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1188.990642] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-9495a91e-d90a-4c5b-ba81-2200f8356cb8 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Deleting contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1188.990874] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-9495a91e-d90a-4c5b-ba81-2200f8356cb8 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Deleting the datastore file [datastore1] 3fa11db4-2f70-495e-a562-ac5f06153847 {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1188.991205] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2344e816-9dad-42bc-bc51-5df7b92ed8f6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.998515] env[63088]: DEBUG oslo_vmware.api [None req-9495a91e-d90a-4c5b-ba81-2200f8356cb8 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Waiting for the task: (returnval){ [ 1188.998515] env[63088]: value = "task-1285589" [ 1188.998515] env[63088]: _type = "Task" [ 1188.998515] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1189.007996] env[63088]: DEBUG oslo_vmware.api [None req-9495a91e-d90a-4c5b-ba81-2200f8356cb8 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285589, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1189.340380] env[63088]: DEBUG oslo_concurrency.lockutils [None req-87ae60bb-0797-4d3c-b527-2195ff31f8b9 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquiring lock "f721d879-55a5-42d1-99c5-871d2e9103ec" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1189.340797] env[63088]: DEBUG oslo_concurrency.lockutils [None req-87ae60bb-0797-4d3c-b527-2195ff31f8b9 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lock "f721d879-55a5-42d1-99c5-871d2e9103ec" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1189.508457] env[63088]: DEBUG oslo_vmware.api [None req-9495a91e-d90a-4c5b-ba81-2200f8356cb8 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Task: {'id': task-1285589, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.134665} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1189.508813] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-9495a91e-d90a-4c5b-ba81-2200f8356cb8 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1189.508860] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-9495a91e-d90a-4c5b-ba81-2200f8356cb8 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Deleted contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1189.509022] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-9495a91e-d90a-4c5b-ba81-2200f8356cb8 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1189.509209] env[63088]: INFO nova.compute.manager [None req-9495a91e-d90a-4c5b-ba81-2200f8356cb8 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1189.509462] env[63088]: DEBUG oslo.service.loopingcall [None req-9495a91e-d90a-4c5b-ba81-2200f8356cb8 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1189.509664] env[63088]: DEBUG nova.compute.manager [-] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1189.509760] env[63088]: DEBUG nova.network.neutron [-] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1189.723994] env[63088]: DEBUG nova.objects.instance [None req-cfae346e-a387-4599-a4b9-6ce6375dfb56 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lazy-loading 'flavor' on Instance uuid 7d8e3c6f-0328-405e-a869-d014caad8641 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1189.843951] env[63088]: DEBUG nova.compute.utils [None req-87ae60bb-0797-4d3c-b527-2195ff31f8b9 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1189.963223] env[63088]: DEBUG nova.compute.manager [req-abacc7d6-3bd3-4041-b462-73d32c8c9161 req-30fa3be1-1e22-446b-a0a7-c54ac7158f3a service nova] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Received event network-vif-deleted-eb4d6233-2d9d-4756-8b9d-084bdfaeb3f7 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1189.963410] env[63088]: INFO nova.compute.manager [req-abacc7d6-3bd3-4041-b462-73d32c8c9161 req-30fa3be1-1e22-446b-a0a7-c54ac7158f3a service nova] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Neutron deleted interface eb4d6233-2d9d-4756-8b9d-084bdfaeb3f7; detaching it from the instance and deleting it from the info cache [ 1189.963617] env[63088]: DEBUG nova.network.neutron [req-abacc7d6-3bd3-4041-b462-73d32c8c9161 req-30fa3be1-1e22-446b-a0a7-c54ac7158f3a service nova] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1190.229187] env[63088]: DEBUG oslo_concurrency.lockutils [None req-cfae346e-a387-4599-a4b9-6ce6375dfb56 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquiring lock "refresh_cache-7d8e3c6f-0328-405e-a869-d014caad8641" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1190.229373] env[63088]: DEBUG oslo_concurrency.lockutils [None req-cfae346e-a387-4599-a4b9-6ce6375dfb56 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquired lock "refresh_cache-7d8e3c6f-0328-405e-a869-d014caad8641" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1190.229552] env[63088]: DEBUG nova.network.neutron [None req-cfae346e-a387-4599-a4b9-6ce6375dfb56 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1190.229759] env[63088]: DEBUG nova.objects.instance [None req-cfae346e-a387-4599-a4b9-6ce6375dfb56 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lazy-loading 'info_cache' on Instance uuid 7d8e3c6f-0328-405e-a869-d014caad8641 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1190.348314] env[63088]: DEBUG oslo_concurrency.lockutils [None req-87ae60bb-0797-4d3c-b527-2195ff31f8b9 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lock "f721d879-55a5-42d1-99c5-871d2e9103ec" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1190.416013] env[63088]: DEBUG nova.network.neutron [-] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1190.466388] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e77576b4-5a11-4721-8309-ca20f0e07438 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.476027] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c106b7a2-0556-4421-8117-53275da95b6b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.498667] env[63088]: DEBUG nova.compute.manager [req-abacc7d6-3bd3-4041-b462-73d32c8c9161 req-30fa3be1-1e22-446b-a0a7-c54ac7158f3a service nova] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Detach interface failed, port_id=eb4d6233-2d9d-4756-8b9d-084bdfaeb3f7, reason: Instance 3fa11db4-2f70-495e-a562-ac5f06153847 could not be found. {{(pid=63088) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1190.736974] env[63088]: DEBUG nova.objects.base [None req-cfae346e-a387-4599-a4b9-6ce6375dfb56 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Object Instance<7d8e3c6f-0328-405e-a869-d014caad8641> lazy-loaded attributes: flavor,info_cache {{(pid=63088) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1190.918392] env[63088]: INFO nova.compute.manager [-] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Took 1.41 seconds to deallocate network for instance. [ 1191.411834] env[63088]: DEBUG oslo_concurrency.lockutils [None req-87ae60bb-0797-4d3c-b527-2195ff31f8b9 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquiring lock "f721d879-55a5-42d1-99c5-871d2e9103ec" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1191.412112] env[63088]: DEBUG oslo_concurrency.lockutils [None req-87ae60bb-0797-4d3c-b527-2195ff31f8b9 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lock "f721d879-55a5-42d1-99c5-871d2e9103ec" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1191.412371] env[63088]: INFO nova.compute.manager [None req-87ae60bb-0797-4d3c-b527-2195ff31f8b9 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Attaching volume 713f8e10-f238-4677-942d-cd09feaf27c7 to /dev/sdb [ 1191.426159] env[63088]: DEBUG oslo_concurrency.lockutils [None req-9495a91e-d90a-4c5b-ba81-2200f8356cb8 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1191.426399] env[63088]: DEBUG oslo_concurrency.lockutils [None req-9495a91e-d90a-4c5b-ba81-2200f8356cb8 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1191.426615] env[63088]: DEBUG nova.objects.instance [None req-9495a91e-d90a-4c5b-ba81-2200f8356cb8 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Lazy-loading 'resources' on Instance uuid 3fa11db4-2f70-495e-a562-ac5f06153847 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1191.443164] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-477f1379-d9e4-4227-b5c4-7b4d7beccea7 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.450224] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec085205-d2c4-4232-af2c-e96f055f3ba3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.464278] env[63088]: DEBUG nova.virt.block_device [None req-87ae60bb-0797-4d3c-b527-2195ff31f8b9 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Updating existing volume attachment record: 75f1574b-c2ff-43d5-9bf8-5db459c6e753 {{(pid=63088) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1191.477398] env[63088]: DEBUG nova.network.neutron [None req-cfae346e-a387-4599-a4b9-6ce6375dfb56 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Updating instance_info_cache with network_info: [{"id": "3e8c16d2-05c9-4554-8a2b-5cdb9b4b3a81", "address": "fa:16:3e:af:74:ac", "network": {"id": "dfcbe153-bc01-4362-9247-de9f0b1c847b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-841757397-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.152", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dace8b5181b84623b08f903d12dfd31e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c68b7663-4f0e-47f0-ac7f-40c6d952f7bb", "external-id": "nsx-vlan-transportzone-696", "segmentation_id": 696, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3e8c16d2-05", "ovs_interfaceid": "3e8c16d2-05c9-4554-8a2b-5cdb9b4b3a81", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1191.947173] env[63088]: DEBUG nova.scheduler.client.report [None req-9495a91e-d90a-4c5b-ba81-2200f8356cb8 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Refreshing inventories for resource provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1191.964223] env[63088]: DEBUG nova.scheduler.client.report [None req-9495a91e-d90a-4c5b-ba81-2200f8356cb8 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Updating ProviderTree inventory for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1191.964223] env[63088]: DEBUG nova.compute.provider_tree [None req-9495a91e-d90a-4c5b-ba81-2200f8356cb8 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Updating inventory in ProviderTree for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1191.974964] env[63088]: DEBUG nova.scheduler.client.report [None req-9495a91e-d90a-4c5b-ba81-2200f8356cb8 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Refreshing aggregate associations for resource provider 6eae54a9-8831-40eb-bf54-4bc60d346b02, aggregates: None {{(pid=63088) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1191.979383] env[63088]: DEBUG oslo_concurrency.lockutils [None req-cfae346e-a387-4599-a4b9-6ce6375dfb56 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Releasing lock "refresh_cache-7d8e3c6f-0328-405e-a869-d014caad8641" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1191.993919] env[63088]: DEBUG nova.scheduler.client.report [None req-9495a91e-d90a-4c5b-ba81-2200f8356cb8 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Refreshing trait associations for resource provider 6eae54a9-8831-40eb-bf54-4bc60d346b02, traits: COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO {{(pid=63088) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1192.041013] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d0b8013-69af-493b-bcd5-bc542cd00944 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.047959] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56f43c09-6b49-46b9-a42a-7604ab96d51f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.079085] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6f4c7ee-bbd9-46af-b1d9-bde13d97c824 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.086713] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cba88dbe-f1b0-4cfc-9868-7f9eba3d82f0 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.099936] env[63088]: DEBUG nova.compute.provider_tree [None req-9495a91e-d90a-4c5b-ba81-2200f8356cb8 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1192.482762] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-cfae346e-a387-4599-a4b9-6ce6375dfb56 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1192.483128] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-caa2f4ee-7f3d-4914-b514-75aa7c3fc1cc {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.490714] env[63088]: DEBUG oslo_vmware.api [None req-cfae346e-a387-4599-a4b9-6ce6375dfb56 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for the task: (returnval){ [ 1192.490714] env[63088]: value = "task-1285593" [ 1192.490714] env[63088]: _type = "Task" [ 1192.490714] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1192.498134] env[63088]: DEBUG oslo_vmware.api [None req-cfae346e-a387-4599-a4b9-6ce6375dfb56 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285593, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1192.602690] env[63088]: DEBUG nova.scheduler.client.report [None req-9495a91e-d90a-4c5b-ba81-2200f8356cb8 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1193.002888] env[63088]: DEBUG oslo_vmware.api [None req-cfae346e-a387-4599-a4b9-6ce6375dfb56 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285593, 'name': PowerOnVM_Task, 'duration_secs': 0.410068} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1193.003208] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-cfae346e-a387-4599-a4b9-6ce6375dfb56 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1193.003395] env[63088]: DEBUG nova.compute.manager [None req-cfae346e-a387-4599-a4b9-6ce6375dfb56 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1193.004259] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d41fc1f6-aa96-4ca3-ac72-6a7629d8efd8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.107347] env[63088]: DEBUG oslo_concurrency.lockutils [None req-9495a91e-d90a-4c5b-ba81-2200f8356cb8 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.681s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1193.127900] env[63088]: INFO nova.scheduler.client.report [None req-9495a91e-d90a-4c5b-ba81-2200f8356cb8 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Deleted allocations for instance 3fa11db4-2f70-495e-a562-ac5f06153847 [ 1193.635607] env[63088]: DEBUG oslo_concurrency.lockutils [None req-9495a91e-d90a-4c5b-ba81-2200f8356cb8 tempest-AttachVolumeShelveTestJSON-937069637 tempest-AttachVolumeShelveTestJSON-937069637-project-member] Lock "3fa11db4-2f70-495e-a562-ac5f06153847" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.233s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1194.748842] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-260da253-ecc1-4502-9dc0-1dc6c61c8e05 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.755611] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-e08def3b-9842-4a22-8206-08753834c310 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Suspending the VM {{(pid=63088) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1163}} [ 1194.755873] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-af1c880d-7a9b-4521-83f0-0a294158721a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.762975] env[63088]: DEBUG oslo_vmware.api [None req-e08def3b-9842-4a22-8206-08753834c310 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for the task: (returnval){ [ 1194.762975] env[63088]: value = "task-1285595" [ 1194.762975] env[63088]: _type = "Task" [ 1194.762975] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1194.770590] env[63088]: DEBUG oslo_vmware.api [None req-e08def3b-9842-4a22-8206-08753834c310 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285595, 'name': SuspendVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1195.272908] env[63088]: DEBUG oslo_vmware.api [None req-e08def3b-9842-4a22-8206-08753834c310 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285595, 'name': SuspendVM_Task} progress is 70%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1195.773506] env[63088]: DEBUG oslo_vmware.api [None req-e08def3b-9842-4a22-8206-08753834c310 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285595, 'name': SuspendVM_Task, 'duration_secs': 0.588619} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1195.773836] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-e08def3b-9842-4a22-8206-08753834c310 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Suspended the VM {{(pid=63088) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1167}} [ 1195.774015] env[63088]: DEBUG nova.compute.manager [None req-e08def3b-9842-4a22-8206-08753834c310 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1195.774777] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-710d9add-ab0b-4d6c-9145-a115d0457ba0 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.006356] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-87ae60bb-0797-4d3c-b527-2195ff31f8b9 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Volume attach. Driver type: vmdk {{(pid=63088) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1196.006607] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-87ae60bb-0797-4d3c-b527-2195ff31f8b9 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-275971', 'volume_id': '713f8e10-f238-4677-942d-cd09feaf27c7', 'name': 'volume-713f8e10-f238-4677-942d-cd09feaf27c7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'f721d879-55a5-42d1-99c5-871d2e9103ec', 'attached_at': '', 'detached_at': '', 'volume_id': '713f8e10-f238-4677-942d-cd09feaf27c7', 'serial': '713f8e10-f238-4677-942d-cd09feaf27c7'} {{(pid=63088) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1196.007501] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85e9133d-cd6d-4fd5-9542-973b78bfd458 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.023248] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c894284-a649-43fa-8f52-13927def61cd {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.047841] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-87ae60bb-0797-4d3c-b527-2195ff31f8b9 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Reconfiguring VM instance instance-0000006a to attach disk [datastore2] volume-713f8e10-f238-4677-942d-cd09feaf27c7/volume-713f8e10-f238-4677-942d-cd09feaf27c7.vmdk or device None with type thin {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1196.048087] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0830d620-c9f9-453f-beea-4a09aa5de0da {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.065652] env[63088]: DEBUG oslo_vmware.api [None req-87ae60bb-0797-4d3c-b527-2195ff31f8b9 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1196.065652] env[63088]: value = "task-1285597" [ 1196.065652] env[63088]: _type = "Task" [ 1196.065652] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1196.073117] env[63088]: DEBUG oslo_vmware.api [None req-87ae60bb-0797-4d3c-b527-2195ff31f8b9 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285597, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1196.578820] env[63088]: DEBUG oslo_vmware.api [None req-87ae60bb-0797-4d3c-b527-2195ff31f8b9 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285597, 'name': ReconfigVM_Task, 'duration_secs': 0.352031} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1196.579163] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-87ae60bb-0797-4d3c-b527-2195ff31f8b9 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Reconfigured VM instance instance-0000006a to attach disk [datastore2] volume-713f8e10-f238-4677-942d-cd09feaf27c7/volume-713f8e10-f238-4677-942d-cd09feaf27c7.vmdk or device None with type thin {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1196.584442] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6dd60b25-1895-4f13-a145-1d08c49eed4f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.599510] env[63088]: DEBUG oslo_vmware.api [None req-87ae60bb-0797-4d3c-b527-2195ff31f8b9 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1196.599510] env[63088]: value = "task-1285598" [ 1196.599510] env[63088]: _type = "Task" [ 1196.599510] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1196.607689] env[63088]: DEBUG oslo_vmware.api [None req-87ae60bb-0797-4d3c-b527-2195ff31f8b9 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285598, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1197.076055] env[63088]: INFO nova.compute.manager [None req-f9fe159f-57b2-4897-bf6f-b266ae0a7e72 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Resuming [ 1197.076414] env[63088]: DEBUG nova.objects.instance [None req-f9fe159f-57b2-4897-bf6f-b266ae0a7e72 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lazy-loading 'flavor' on Instance uuid 7d8e3c6f-0328-405e-a869-d014caad8641 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1197.109253] env[63088]: DEBUG oslo_vmware.api [None req-87ae60bb-0797-4d3c-b527-2195ff31f8b9 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285598, 'name': ReconfigVM_Task, 'duration_secs': 0.137897} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1197.109543] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-87ae60bb-0797-4d3c-b527-2195ff31f8b9 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-275971', 'volume_id': '713f8e10-f238-4677-942d-cd09feaf27c7', 'name': 'volume-713f8e10-f238-4677-942d-cd09feaf27c7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'f721d879-55a5-42d1-99c5-871d2e9103ec', 'attached_at': '', 'detached_at': '', 'volume_id': '713f8e10-f238-4677-942d-cd09feaf27c7', 'serial': '713f8e10-f238-4677-942d-cd09feaf27c7'} {{(pid=63088) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1198.084730] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f9fe159f-57b2-4897-bf6f-b266ae0a7e72 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquiring lock "refresh_cache-7d8e3c6f-0328-405e-a869-d014caad8641" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1198.085123] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f9fe159f-57b2-4897-bf6f-b266ae0a7e72 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquired lock "refresh_cache-7d8e3c6f-0328-405e-a869-d014caad8641" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1198.085160] env[63088]: DEBUG nova.network.neutron [None req-f9fe159f-57b2-4897-bf6f-b266ae0a7e72 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1198.142247] env[63088]: DEBUG nova.objects.instance [None req-87ae60bb-0797-4d3c-b527-2195ff31f8b9 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lazy-loading 'flavor' on Instance uuid f721d879-55a5-42d1-99c5-871d2e9103ec {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1198.647295] env[63088]: DEBUG oslo_concurrency.lockutils [None req-87ae60bb-0797-4d3c-b527-2195ff31f8b9 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lock "f721d879-55a5-42d1-99c5-871d2e9103ec" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.235s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1198.873311] env[63088]: DEBUG nova.network.neutron [None req-f9fe159f-57b2-4897-bf6f-b266ae0a7e72 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Updating instance_info_cache with network_info: [{"id": "3e8c16d2-05c9-4554-8a2b-5cdb9b4b3a81", "address": "fa:16:3e:af:74:ac", "network": {"id": "dfcbe153-bc01-4362-9247-de9f0b1c847b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-841757397-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.152", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dace8b5181b84623b08f903d12dfd31e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c68b7663-4f0e-47f0-ac7f-40c6d952f7bb", "external-id": "nsx-vlan-transportzone-696", "segmentation_id": 696, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3e8c16d2-05", "ovs_interfaceid": "3e8c16d2-05c9-4554-8a2b-5cdb9b4b3a81", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1199.376205] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f9fe159f-57b2-4897-bf6f-b266ae0a7e72 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Releasing lock "refresh_cache-7d8e3c6f-0328-405e-a869-d014caad8641" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1199.377198] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a29a9930-5672-416d-b1ea-39a15848e76d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.384578] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-f9fe159f-57b2-4897-bf6f-b266ae0a7e72 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Resuming the VM {{(pid=63088) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1184}} [ 1199.384890] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2021d2ec-2520-48b2-a07b-06bc1a18b787 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.391092] env[63088]: DEBUG oslo_vmware.api [None req-f9fe159f-57b2-4897-bf6f-b266ae0a7e72 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for the task: (returnval){ [ 1199.391092] env[63088]: value = "task-1285599" [ 1199.391092] env[63088]: _type = "Task" [ 1199.391092] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1199.398745] env[63088]: DEBUG oslo_vmware.api [None req-f9fe159f-57b2-4897-bf6f-b266ae0a7e72 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285599, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1199.507780] env[63088]: DEBUG nova.compute.manager [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Stashing vm_state: active {{(pid=63088) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1199.901637] env[63088]: DEBUG oslo_vmware.api [None req-f9fe159f-57b2-4897-bf6f-b266ae0a7e72 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285599, 'name': PowerOnVM_Task, 'duration_secs': 0.498596} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1199.901988] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-f9fe159f-57b2-4897-bf6f-b266ae0a7e72 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Resumed the VM {{(pid=63088) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1189}} [ 1199.902213] env[63088]: DEBUG nova.compute.manager [None req-f9fe159f-57b2-4897-bf6f-b266ae0a7e72 tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1199.903127] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-435fe984-afdc-4214-9ec3-eb7a079e3c85 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.031814] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1200.032176] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1200.536789] env[63088]: INFO nova.compute.claims [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1200.900987] env[63088]: DEBUG oslo_concurrency.lockutils [None req-61a1e709-e313-4740-ac97-62a883601b3d tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquiring lock "7d8e3c6f-0328-405e-a869-d014caad8641" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1200.901316] env[63088]: DEBUG oslo_concurrency.lockutils [None req-61a1e709-e313-4740-ac97-62a883601b3d tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lock "7d8e3c6f-0328-405e-a869-d014caad8641" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1200.901534] env[63088]: DEBUG oslo_concurrency.lockutils [None req-61a1e709-e313-4740-ac97-62a883601b3d tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquiring lock "7d8e3c6f-0328-405e-a869-d014caad8641-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1200.901729] env[63088]: DEBUG oslo_concurrency.lockutils [None req-61a1e709-e313-4740-ac97-62a883601b3d tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lock "7d8e3c6f-0328-405e-a869-d014caad8641-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1200.901904] env[63088]: DEBUG oslo_concurrency.lockutils [None req-61a1e709-e313-4740-ac97-62a883601b3d tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lock "7d8e3c6f-0328-405e-a869-d014caad8641-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1200.904096] env[63088]: INFO nova.compute.manager [None req-61a1e709-e313-4740-ac97-62a883601b3d tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Terminating instance [ 1200.905926] env[63088]: DEBUG nova.compute.manager [None req-61a1e709-e313-4740-ac97-62a883601b3d tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1200.906137] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-61a1e709-e313-4740-ac97-62a883601b3d tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1200.906963] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-796c7233-a4db-48d6-a67b-36a1fc2f8ae1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.915919] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-61a1e709-e313-4740-ac97-62a883601b3d tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1200.916164] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b8aec966-eed5-434c-b79b-de470592efc9 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.921918] env[63088]: DEBUG oslo_vmware.api [None req-61a1e709-e313-4740-ac97-62a883601b3d tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for the task: (returnval){ [ 1200.921918] env[63088]: value = "task-1285600" [ 1200.921918] env[63088]: _type = "Task" [ 1200.921918] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1200.929418] env[63088]: DEBUG oslo_vmware.api [None req-61a1e709-e313-4740-ac97-62a883601b3d tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285600, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1201.042502] env[63088]: INFO nova.compute.resource_tracker [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Updating resource usage from migration 1047a6e9-8d90-4ad4-b322-8fe7c2def26c [ 1201.093778] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74015d21-3c76-45d2-8503-bb8a5b835889 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.102559] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e4b4222-eea3-42ff-aed5-b1937c12b4be {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.131204] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bc67828-a35f-46fd-8784-26893ec35162 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.137889] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0f062fe-6d9d-4cf3-8b82-2a3e9d471035 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.151732] env[63088]: DEBUG nova.compute.provider_tree [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1201.431907] env[63088]: DEBUG oslo_vmware.api [None req-61a1e709-e313-4740-ac97-62a883601b3d tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285600, 'name': PowerOffVM_Task, 'duration_secs': 0.208535} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1201.432219] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-61a1e709-e313-4740-ac97-62a883601b3d tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1201.432422] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-61a1e709-e313-4740-ac97-62a883601b3d tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Unregistering the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1201.432667] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d5bd36b1-9402-46c3-94c2-73370bae7391 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.491983] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-61a1e709-e313-4740-ac97-62a883601b3d tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Unregistered the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1201.492223] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-61a1e709-e313-4740-ac97-62a883601b3d tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Deleting contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1201.492412] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-61a1e709-e313-4740-ac97-62a883601b3d tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Deleting the datastore file [datastore1] 7d8e3c6f-0328-405e-a869-d014caad8641 {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1201.492670] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ae3914e5-6102-4560-a977-f3c6bc9d3e7e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.498737] env[63088]: DEBUG oslo_vmware.api [None req-61a1e709-e313-4740-ac97-62a883601b3d tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for the task: (returnval){ [ 1201.498737] env[63088]: value = "task-1285602" [ 1201.498737] env[63088]: _type = "Task" [ 1201.498737] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1201.506537] env[63088]: DEBUG oslo_vmware.api [None req-61a1e709-e313-4740-ac97-62a883601b3d tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285602, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1201.655099] env[63088]: DEBUG nova.scheduler.client.report [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1202.008539] env[63088]: DEBUG oslo_vmware.api [None req-61a1e709-e313-4740-ac97-62a883601b3d tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Task: {'id': task-1285602, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.147888} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1202.008728] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-61a1e709-e313-4740-ac97-62a883601b3d tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1202.008919] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-61a1e709-e313-4740-ac97-62a883601b3d tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Deleted contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1202.009114] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-61a1e709-e313-4740-ac97-62a883601b3d tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1202.009371] env[63088]: INFO nova.compute.manager [None req-61a1e709-e313-4740-ac97-62a883601b3d tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1202.009634] env[63088]: DEBUG oslo.service.loopingcall [None req-61a1e709-e313-4740-ac97-62a883601b3d tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1202.009839] env[63088]: DEBUG nova.compute.manager [-] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1202.009933] env[63088]: DEBUG nova.network.neutron [-] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1202.160092] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.128s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1202.160318] env[63088]: INFO nova.compute.manager [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Migrating [ 1202.417739] env[63088]: DEBUG nova.compute.manager [req-45a8caa9-69f3-437a-add8-0e7a3d48b4bc req-78923979-a4dc-44f7-b837-b91725c2e484 service nova] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Received event network-vif-deleted-3e8c16d2-05c9-4554-8a2b-5cdb9b4b3a81 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1202.417997] env[63088]: INFO nova.compute.manager [req-45a8caa9-69f3-437a-add8-0e7a3d48b4bc req-78923979-a4dc-44f7-b837-b91725c2e484 service nova] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Neutron deleted interface 3e8c16d2-05c9-4554-8a2b-5cdb9b4b3a81; detaching it from the instance and deleting it from the info cache [ 1202.418177] env[63088]: DEBUG nova.network.neutron [req-45a8caa9-69f3-437a-add8-0e7a3d48b4bc req-78923979-a4dc-44f7-b837-b91725c2e484 service nova] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1202.674385] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquiring lock "refresh_cache-f721d879-55a5-42d1-99c5-871d2e9103ec" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1202.674671] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquired lock "refresh_cache-f721d879-55a5-42d1-99c5-871d2e9103ec" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1202.674769] env[63088]: DEBUG nova.network.neutron [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1202.891197] env[63088]: DEBUG nova.network.neutron [-] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1202.920258] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-31f8974a-90ae-4753-9d18-b7bd51536912 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.931715] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c091aea6-3012-47f4-9b51-4cdf34abcd8a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.955684] env[63088]: DEBUG nova.compute.manager [req-45a8caa9-69f3-437a-add8-0e7a3d48b4bc req-78923979-a4dc-44f7-b837-b91725c2e484 service nova] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Detach interface failed, port_id=3e8c16d2-05c9-4554-8a2b-5cdb9b4b3a81, reason: Instance 7d8e3c6f-0328-405e-a869-d014caad8641 could not be found. {{(pid=63088) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1203.383195] env[63088]: DEBUG nova.network.neutron [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Updating instance_info_cache with network_info: [{"id": "154da9f6-d230-479a-88e0-f13ebc655493", "address": "fa:16:3e:47:57:05", "network": {"id": "711d3d9a-86db-4aa1-87b5-481f1812dfa4", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-59376186-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.225", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a7898ddafe0d41038a6ae4277f2c2f48", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e41070eb-3ac1-4ca9-a3d0-fd65893a97de", "external-id": "nsx-vlan-transportzone-596", "segmentation_id": 596, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap154da9f6-d2", "ovs_interfaceid": "154da9f6-d230-479a-88e0-f13ebc655493", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1203.394555] env[63088]: INFO nova.compute.manager [-] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Took 1.38 seconds to deallocate network for instance. [ 1203.886476] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Releasing lock "refresh_cache-f721d879-55a5-42d1-99c5-871d2e9103ec" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1203.900107] env[63088]: DEBUG oslo_concurrency.lockutils [None req-61a1e709-e313-4740-ac97-62a883601b3d tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1203.900382] env[63088]: DEBUG oslo_concurrency.lockutils [None req-61a1e709-e313-4740-ac97-62a883601b3d tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1203.900604] env[63088]: DEBUG nova.objects.instance [None req-61a1e709-e313-4740-ac97-62a883601b3d tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lazy-loading 'resources' on Instance uuid 7d8e3c6f-0328-405e-a869-d014caad8641 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1204.454265] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75d24870-a444-4361-b285-26da86e2f30e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.462142] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c38e2da9-d910-4f8d-ba4e-f6e9abfda745 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.491396] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-923537fa-17e9-4497-8e42-5b6e3de8ad63 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.498403] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15496228-6460-4d6d-bbf0-e2570dd18290 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.511389] env[63088]: DEBUG nova.compute.provider_tree [None req-61a1e709-e313-4740-ac97-62a883601b3d tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1205.015380] env[63088]: DEBUG nova.scheduler.client.report [None req-61a1e709-e313-4740-ac97-62a883601b3d tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1205.403020] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b37c5fb-09f1-4fd8-9852-1e44379c715f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.424854] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Updating instance 'f721d879-55a5-42d1-99c5-871d2e9103ec' progress to 0 {{(pid=63088) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1205.519558] env[63088]: DEBUG oslo_concurrency.lockutils [None req-61a1e709-e313-4740-ac97-62a883601b3d tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.619s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1205.537306] env[63088]: INFO nova.scheduler.client.report [None req-61a1e709-e313-4740-ac97-62a883601b3d tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Deleted allocations for instance 7d8e3c6f-0328-405e-a869-d014caad8641 [ 1205.931048] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1205.931383] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-43d7aba1-8a26-45f9-9121-6b740e9b8a76 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.939028] env[63088]: DEBUG oslo_vmware.api [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1205.939028] env[63088]: value = "task-1285603" [ 1205.939028] env[63088]: _type = "Task" [ 1205.939028] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1205.947701] env[63088]: DEBUG oslo_vmware.api [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285603, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1206.044282] env[63088]: DEBUG oslo_concurrency.lockutils [None req-61a1e709-e313-4740-ac97-62a883601b3d tempest-ServerActionsTestJSON-2019694924 tempest-ServerActionsTestJSON-2019694924-project-member] Lock "7d8e3c6f-0328-405e-a869-d014caad8641" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.143s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1206.449535] env[63088]: DEBUG oslo_vmware.api [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285603, 'name': PowerOffVM_Task, 'duration_secs': 0.181655} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1206.449800] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1206.449991] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Updating instance 'f721d879-55a5-42d1-99c5-871d2e9103ec' progress to 17 {{(pid=63088) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1206.957635] env[63088]: DEBUG nova.virt.hardware [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:33Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1206.957909] env[63088]: DEBUG nova.virt.hardware [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1206.958221] env[63088]: DEBUG nova.virt.hardware [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1206.958318] env[63088]: DEBUG nova.virt.hardware [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1206.958389] env[63088]: DEBUG nova.virt.hardware [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1206.958539] env[63088]: DEBUG nova.virt.hardware [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1206.958751] env[63088]: DEBUG nova.virt.hardware [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1206.958914] env[63088]: DEBUG nova.virt.hardware [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1206.959094] env[63088]: DEBUG nova.virt.hardware [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1206.959269] env[63088]: DEBUG nova.virt.hardware [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1206.959444] env[63088]: DEBUG nova.virt.hardware [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1206.964670] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1debb010-3ff7-4283-aa41-42383e06ce79 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.980961] env[63088]: DEBUG oslo_vmware.api [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1206.980961] env[63088]: value = "task-1285604" [ 1206.980961] env[63088]: _type = "Task" [ 1206.980961] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1206.991562] env[63088]: DEBUG oslo_vmware.api [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285604, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1207.491841] env[63088]: DEBUG oslo_vmware.api [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285604, 'name': ReconfigVM_Task, 'duration_secs': 0.162907} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1207.492280] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Updating instance 'f721d879-55a5-42d1-99c5-871d2e9103ec' progress to 33 {{(pid=63088) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1207.999584] env[63088]: DEBUG nova.virt.hardware [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1207.999890] env[63088]: DEBUG nova.virt.hardware [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1208.000123] env[63088]: DEBUG nova.virt.hardware [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1208.000365] env[63088]: DEBUG nova.virt.hardware [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1208.000550] env[63088]: DEBUG nova.virt.hardware [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1208.000773] env[63088]: DEBUG nova.virt.hardware [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1208.001135] env[63088]: DEBUG nova.virt.hardware [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1208.001135] env[63088]: DEBUG nova.virt.hardware [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1208.001295] env[63088]: DEBUG nova.virt.hardware [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1208.001464] env[63088]: DEBUG nova.virt.hardware [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1208.001756] env[63088]: DEBUG nova.virt.hardware [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1208.007100] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Reconfiguring VM instance instance-0000006a to detach disk 2000 {{(pid=63088) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1208.007424] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cdd83424-ceca-486b-8335-62a8b23678e0 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.027329] env[63088]: DEBUG oslo_vmware.api [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1208.027329] env[63088]: value = "task-1285605" [ 1208.027329] env[63088]: _type = "Task" [ 1208.027329] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1208.035816] env[63088]: DEBUG oslo_vmware.api [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285605, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1208.540905] env[63088]: DEBUG oslo_vmware.api [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285605, 'name': ReconfigVM_Task, 'duration_secs': 0.193351} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1208.540905] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Reconfigured VM instance instance-0000006a to detach disk 2000 {{(pid=63088) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1208.540905] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd4e8190-5622-4cf3-a5d5-3d4a32722f5e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.564054] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Reconfiguring VM instance instance-0000006a to attach disk [datastore1] f721d879-55a5-42d1-99c5-871d2e9103ec/f721d879-55a5-42d1-99c5-871d2e9103ec.vmdk or device None with type thin {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1208.564295] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cd00c63b-6f33-4d15-93b5-93c4e756e52d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.583062] env[63088]: DEBUG oslo_vmware.api [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1208.583062] env[63088]: value = "task-1285606" [ 1208.583062] env[63088]: _type = "Task" [ 1208.583062] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1208.590483] env[63088]: DEBUG oslo_vmware.api [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285606, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1209.092941] env[63088]: DEBUG oslo_vmware.api [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285606, 'name': ReconfigVM_Task, 'duration_secs': 0.359324} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1209.093242] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Reconfigured VM instance instance-0000006a to attach disk [datastore1] f721d879-55a5-42d1-99c5-871d2e9103ec/f721d879-55a5-42d1-99c5-871d2e9103ec.vmdk or device None with type thin {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1209.093529] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Updating instance 'f721d879-55a5-42d1-99c5-871d2e9103ec' progress to 50 {{(pid=63088) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1209.600881] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26474ea3-e4af-4c5f-8711-71ee02f59f5a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.624868] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2eb9c4d-fefd-4d71-bd4b-ccd4c094b79c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.645943] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Updating instance 'f721d879-55a5-42d1-99c5-871d2e9103ec' progress to 67 {{(pid=63088) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1210.967936] env[63088]: DEBUG oslo_concurrency.lockutils [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Acquiring lock "8e1bb264-bd9a-4779-9bbf-849c14c650a9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1210.968250] env[63088]: DEBUG oslo_concurrency.lockutils [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Lock "8e1bb264-bd9a-4779-9bbf-849c14c650a9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1211.291634] env[63088]: DEBUG nova.network.neutron [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Port 154da9f6-d230-479a-88e0-f13ebc655493 binding to destination host cpu-1 is already ACTIVE {{(pid=63088) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1211.470502] env[63088]: DEBUG nova.compute.manager [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] [instance: 8e1bb264-bd9a-4779-9bbf-849c14c650a9] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1211.994634] env[63088]: DEBUG oslo_concurrency.lockutils [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1211.996021] env[63088]: DEBUG oslo_concurrency.lockutils [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1211.996785] env[63088]: INFO nova.compute.claims [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] [instance: 8e1bb264-bd9a-4779-9bbf-849c14c650a9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1212.314301] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquiring lock "f721d879-55a5-42d1-99c5-871d2e9103ec-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1212.314501] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lock "f721d879-55a5-42d1-99c5-871d2e9103ec-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1212.314677] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lock "f721d879-55a5-42d1-99c5-871d2e9103ec-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1213.049201] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5a3f7b7-2a78-40e0-a7a0-57a24cc06b84 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.056913] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10475e69-3a12-47d8-9172-761ba1e15c4c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.087114] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de6501bb-1e80-451c-aa8f-cec2a5d3008e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.093549] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c43fa9fc-e6aa-417a-a174-58b71b558f33 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.105940] env[63088]: DEBUG nova.compute.provider_tree [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1213.347974] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquiring lock "refresh_cache-f721d879-55a5-42d1-99c5-871d2e9103ec" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1213.348212] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquired lock "refresh_cache-f721d879-55a5-42d1-99c5-871d2e9103ec" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1213.348396] env[63088]: DEBUG nova.network.neutron [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1213.608743] env[63088]: DEBUG nova.scheduler.client.report [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1214.082527] env[63088]: DEBUG nova.network.neutron [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Updating instance_info_cache with network_info: [{"id": "154da9f6-d230-479a-88e0-f13ebc655493", "address": "fa:16:3e:47:57:05", "network": {"id": "711d3d9a-86db-4aa1-87b5-481f1812dfa4", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-59376186-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.225", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a7898ddafe0d41038a6ae4277f2c2f48", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e41070eb-3ac1-4ca9-a3d0-fd65893a97de", "external-id": "nsx-vlan-transportzone-596", "segmentation_id": 596, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap154da9f6-d2", "ovs_interfaceid": "154da9f6-d230-479a-88e0-f13ebc655493", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1214.113513] env[63088]: DEBUG oslo_concurrency.lockutils [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.118s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1214.114118] env[63088]: DEBUG nova.compute.manager [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] [instance: 8e1bb264-bd9a-4779-9bbf-849c14c650a9] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1214.585245] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Releasing lock "refresh_cache-f721d879-55a5-42d1-99c5-871d2e9103ec" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1214.618393] env[63088]: DEBUG nova.compute.utils [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1214.619749] env[63088]: DEBUG nova.compute.manager [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] [instance: 8e1bb264-bd9a-4779-9bbf-849c14c650a9] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1214.619918] env[63088]: DEBUG nova.network.neutron [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] [instance: 8e1bb264-bd9a-4779-9bbf-849c14c650a9] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1214.656873] env[63088]: DEBUG nova.policy [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e034ff570128429886c3fab6cc1f3ac8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '93772004117748829e0a157827982ac8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 1214.880388] env[63088]: DEBUG nova.network.neutron [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] [instance: 8e1bb264-bd9a-4779-9bbf-849c14c650a9] Successfully created port: de4db792-3bff-4ed9-875e-8bfaeed5d0ab {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1215.095152] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a31ae24a-0207-4081-9b2b-1b5d84e3ff2e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.102687] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81d65247-e7c2-40ab-a5f4-777ecb5c9939 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.122417] env[63088]: DEBUG nova.compute.manager [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] [instance: 8e1bb264-bd9a-4779-9bbf-849c14c650a9] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1216.132296] env[63088]: DEBUG nova.compute.manager [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] [instance: 8e1bb264-bd9a-4779-9bbf-849c14c650a9] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1216.157066] env[63088]: DEBUG nova.virt.hardware [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1216.157318] env[63088]: DEBUG nova.virt.hardware [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1216.157479] env[63088]: DEBUG nova.virt.hardware [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1216.157663] env[63088]: DEBUG nova.virt.hardware [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1216.157815] env[63088]: DEBUG nova.virt.hardware [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1216.157964] env[63088]: DEBUG nova.virt.hardware [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1216.158188] env[63088]: DEBUG nova.virt.hardware [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1216.158356] env[63088]: DEBUG nova.virt.hardware [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1216.158524] env[63088]: DEBUG nova.virt.hardware [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1216.158687] env[63088]: DEBUG nova.virt.hardware [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1216.158859] env[63088]: DEBUG nova.virt.hardware [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1216.159728] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12286830-b31b-415c-9bfd-a04df76fac1c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.168309] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2cbef9e-1a32-449c-adf9-3c16a448d6a0 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.197199] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aeed6f16-169d-4a37-ac05-8cfc36e351dc {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.217643] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eadc0cc1-764d-4475-b6fd-1d6b4bded27b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.224550] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Updating instance 'f721d879-55a5-42d1-99c5-871d2e9103ec' progress to 83 {{(pid=63088) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1216.251999] env[63088]: DEBUG nova.compute.manager [req-24fbcbf2-874f-4b9d-8fe6-32a8cd47489a req-b7369320-4d7d-45e0-b0bb-c58e79389496 service nova] [instance: 8e1bb264-bd9a-4779-9bbf-849c14c650a9] Received event network-vif-plugged-de4db792-3bff-4ed9-875e-8bfaeed5d0ab {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1216.252241] env[63088]: DEBUG oslo_concurrency.lockutils [req-24fbcbf2-874f-4b9d-8fe6-32a8cd47489a req-b7369320-4d7d-45e0-b0bb-c58e79389496 service nova] Acquiring lock "8e1bb264-bd9a-4779-9bbf-849c14c650a9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1216.252463] env[63088]: DEBUG oslo_concurrency.lockutils [req-24fbcbf2-874f-4b9d-8fe6-32a8cd47489a req-b7369320-4d7d-45e0-b0bb-c58e79389496 service nova] Lock "8e1bb264-bd9a-4779-9bbf-849c14c650a9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1216.252667] env[63088]: DEBUG oslo_concurrency.lockutils [req-24fbcbf2-874f-4b9d-8fe6-32a8cd47489a req-b7369320-4d7d-45e0-b0bb-c58e79389496 service nova] Lock "8e1bb264-bd9a-4779-9bbf-849c14c650a9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1216.252778] env[63088]: DEBUG nova.compute.manager [req-24fbcbf2-874f-4b9d-8fe6-32a8cd47489a req-b7369320-4d7d-45e0-b0bb-c58e79389496 service nova] [instance: 8e1bb264-bd9a-4779-9bbf-849c14c650a9] No waiting events found dispatching network-vif-plugged-de4db792-3bff-4ed9-875e-8bfaeed5d0ab {{(pid=63088) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1216.252946] env[63088]: WARNING nova.compute.manager [req-24fbcbf2-874f-4b9d-8fe6-32a8cd47489a req-b7369320-4d7d-45e0-b0bb-c58e79389496 service nova] [instance: 8e1bb264-bd9a-4779-9bbf-849c14c650a9] Received unexpected event network-vif-plugged-de4db792-3bff-4ed9-875e-8bfaeed5d0ab for instance with vm_state building and task_state spawning. [ 1216.373885] env[63088]: DEBUG nova.network.neutron [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] [instance: 8e1bb264-bd9a-4779-9bbf-849c14c650a9] Successfully updated port: de4db792-3bff-4ed9-875e-8bfaeed5d0ab {{(pid=63088) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1216.730783] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1216.731167] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a4970b3d-e194-4776-8db1-edc9b0746d8e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.738604] env[63088]: DEBUG oslo_vmware.api [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1216.738604] env[63088]: value = "task-1285607" [ 1216.738604] env[63088]: _type = "Task" [ 1216.738604] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1216.746042] env[63088]: DEBUG oslo_vmware.api [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285607, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1216.876793] env[63088]: DEBUG oslo_concurrency.lockutils [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Acquiring lock "refresh_cache-8e1bb264-bd9a-4779-9bbf-849c14c650a9" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1216.876793] env[63088]: DEBUG oslo_concurrency.lockutils [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Acquired lock "refresh_cache-8e1bb264-bd9a-4779-9bbf-849c14c650a9" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1216.876793] env[63088]: DEBUG nova.network.neutron [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] [instance: 8e1bb264-bd9a-4779-9bbf-849c14c650a9] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1217.249605] env[63088]: DEBUG oslo_vmware.api [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285607, 'name': PowerOnVM_Task, 'duration_secs': 0.371914} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1217.249605] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1217.249605] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-f5777e9d-2bdd-4d89-8614-52774b587e0e tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Updating instance 'f721d879-55a5-42d1-99c5-871d2e9103ec' progress to 100 {{(pid=63088) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1217.302992] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1217.417596] env[63088]: DEBUG nova.network.neutron [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] [instance: 8e1bb264-bd9a-4779-9bbf-849c14c650a9] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1217.574208] env[63088]: DEBUG nova.network.neutron [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] [instance: 8e1bb264-bd9a-4779-9bbf-849c14c650a9] Updating instance_info_cache with network_info: [{"id": "de4db792-3bff-4ed9-875e-8bfaeed5d0ab", "address": "fa:16:3e:73:50:ea", "network": {"id": "a5517464-78fd-49b1-93ab-33d1d36b1945", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-1988347292-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "93772004117748829e0a157827982ac8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "678ebbe4-4c53-4eaf-a689-93981310f37d", "external-id": "nsx-vlan-transportzone-443", "segmentation_id": 443, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapde4db792-3b", "ovs_interfaceid": "de4db792-3bff-4ed9-875e-8bfaeed5d0ab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1217.603960] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1218.076496] env[63088]: DEBUG oslo_concurrency.lockutils [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Releasing lock "refresh_cache-8e1bb264-bd9a-4779-9bbf-849c14c650a9" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1218.076841] env[63088]: DEBUG nova.compute.manager [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] [instance: 8e1bb264-bd9a-4779-9bbf-849c14c650a9] Instance network_info: |[{"id": "de4db792-3bff-4ed9-875e-8bfaeed5d0ab", "address": "fa:16:3e:73:50:ea", "network": {"id": "a5517464-78fd-49b1-93ab-33d1d36b1945", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-1988347292-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "93772004117748829e0a157827982ac8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "678ebbe4-4c53-4eaf-a689-93981310f37d", "external-id": "nsx-vlan-transportzone-443", "segmentation_id": 443, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapde4db792-3b", "ovs_interfaceid": "de4db792-3bff-4ed9-875e-8bfaeed5d0ab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1218.077358] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] [instance: 8e1bb264-bd9a-4779-9bbf-849c14c650a9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:73:50:ea', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '678ebbe4-4c53-4eaf-a689-93981310f37d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'de4db792-3bff-4ed9-875e-8bfaeed5d0ab', 'vif_model': 'vmxnet3'}] {{(pid=63088) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1218.084645] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Creating folder: Project (93772004117748829e0a157827982ac8). Parent ref: group-v275816. {{(pid=63088) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1218.084912] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8dde09e4-568d-4bd5-8200-295ea1c1f190 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.095894] env[63088]: INFO nova.virt.vmwareapi.vm_util [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Created folder: Project (93772004117748829e0a157827982ac8) in parent group-v275816. [ 1218.096215] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Creating folder: Instances. Parent ref: group-v275972. {{(pid=63088) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1218.096365] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-784d8295-10e2-4ad5-b07d-2ab1607e0fdd {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.107261] env[63088]: INFO nova.virt.vmwareapi.vm_util [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Created folder: Instances in parent group-v275972. [ 1218.107490] env[63088]: DEBUG oslo.service.loopingcall [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1218.107678] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8e1bb264-bd9a-4779-9bbf-849c14c650a9] Creating VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1218.107875] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9bfdc650-8e28-4ffa-ad4e-6b42e8ccf4ab {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.125726] env[63088]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1218.125726] env[63088]: value = "task-1285610" [ 1218.125726] env[63088]: _type = "Task" [ 1218.125726] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1218.132916] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285610, 'name': CreateVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1218.278579] env[63088]: DEBUG nova.compute.manager [req-0e1636e6-36b0-4dbe-a498-f007a880e127 req-7bcbc400-8563-472e-820d-fef7a2a3603e service nova] [instance: 8e1bb264-bd9a-4779-9bbf-849c14c650a9] Received event network-changed-de4db792-3bff-4ed9-875e-8bfaeed5d0ab {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1218.278890] env[63088]: DEBUG nova.compute.manager [req-0e1636e6-36b0-4dbe-a498-f007a880e127 req-7bcbc400-8563-472e-820d-fef7a2a3603e service nova] [instance: 8e1bb264-bd9a-4779-9bbf-849c14c650a9] Refreshing instance network info cache due to event network-changed-de4db792-3bff-4ed9-875e-8bfaeed5d0ab. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1218.279030] env[63088]: DEBUG oslo_concurrency.lockutils [req-0e1636e6-36b0-4dbe-a498-f007a880e127 req-7bcbc400-8563-472e-820d-fef7a2a3603e service nova] Acquiring lock "refresh_cache-8e1bb264-bd9a-4779-9bbf-849c14c650a9" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1218.279173] env[63088]: DEBUG oslo_concurrency.lockutils [req-0e1636e6-36b0-4dbe-a498-f007a880e127 req-7bcbc400-8563-472e-820d-fef7a2a3603e service nova] Acquired lock "refresh_cache-8e1bb264-bd9a-4779-9bbf-849c14c650a9" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1218.279337] env[63088]: DEBUG nova.network.neutron [req-0e1636e6-36b0-4dbe-a498-f007a880e127 req-7bcbc400-8563-472e-820d-fef7a2a3603e service nova] [instance: 8e1bb264-bd9a-4779-9bbf-849c14c650a9] Refreshing network info cache for port de4db792-3bff-4ed9-875e-8bfaeed5d0ab {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1218.603591] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1218.603854] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1218.604835] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63088) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1218.635428] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285610, 'name': CreateVM_Task, 'duration_secs': 0.286539} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1218.635601] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8e1bb264-bd9a-4779-9bbf-849c14c650a9] Created VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1218.636436] env[63088]: DEBUG oslo_concurrency.lockutils [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1218.636624] env[63088]: DEBUG oslo_concurrency.lockutils [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1218.636961] env[63088]: DEBUG oslo_concurrency.lockutils [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1218.637227] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f12422de-5275-4c59-af17-14149903fffc {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.641389] env[63088]: DEBUG oslo_vmware.api [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Waiting for the task: (returnval){ [ 1218.641389] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52a673f3-b90d-f66a-96e3-3c7db7dece12" [ 1218.641389] env[63088]: _type = "Task" [ 1218.641389] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1218.648543] env[63088]: DEBUG oslo_vmware.api [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52a673f3-b90d-f66a-96e3-3c7db7dece12, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1218.963961] env[63088]: DEBUG nova.network.neutron [req-0e1636e6-36b0-4dbe-a498-f007a880e127 req-7bcbc400-8563-472e-820d-fef7a2a3603e service nova] [instance: 8e1bb264-bd9a-4779-9bbf-849c14c650a9] Updated VIF entry in instance network info cache for port de4db792-3bff-4ed9-875e-8bfaeed5d0ab. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1218.964379] env[63088]: DEBUG nova.network.neutron [req-0e1636e6-36b0-4dbe-a498-f007a880e127 req-7bcbc400-8563-472e-820d-fef7a2a3603e service nova] [instance: 8e1bb264-bd9a-4779-9bbf-849c14c650a9] Updating instance_info_cache with network_info: [{"id": "de4db792-3bff-4ed9-875e-8bfaeed5d0ab", "address": "fa:16:3e:73:50:ea", "network": {"id": "a5517464-78fd-49b1-93ab-33d1d36b1945", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-1988347292-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "93772004117748829e0a157827982ac8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "678ebbe4-4c53-4eaf-a689-93981310f37d", "external-id": "nsx-vlan-transportzone-443", "segmentation_id": 443, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapde4db792-3b", "ovs_interfaceid": "de4db792-3bff-4ed9-875e-8bfaeed5d0ab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1219.151477] env[63088]: DEBUG oslo_vmware.api [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52a673f3-b90d-f66a-96e3-3c7db7dece12, 'name': SearchDatastore_Task, 'duration_secs': 0.009708} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1219.151790] env[63088]: DEBUG oslo_concurrency.lockutils [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1219.152040] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] [instance: 8e1bb264-bd9a-4779-9bbf-849c14c650a9] Processing image 1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1219.152318] env[63088]: DEBUG oslo_concurrency.lockutils [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1219.152484] env[63088]: DEBUG oslo_concurrency.lockutils [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1219.152655] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1219.152906] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0517bfc0-5ed1-4f0d-adaa-8da71f94161b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.160191] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1219.160364] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1219.161037] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7fca1a92-7cc9-46b0-8934-1c7ce7f92ce7 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.165526] env[63088]: DEBUG oslo_vmware.api [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Waiting for the task: (returnval){ [ 1219.165526] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52fa3614-9010-bc98-2665-149353ca8d30" [ 1219.165526] env[63088]: _type = "Task" [ 1219.165526] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1219.172714] env[63088]: DEBUG oslo_vmware.api [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52fa3614-9010-bc98-2665-149353ca8d30, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1219.328906] env[63088]: DEBUG nova.network.neutron [None req-e0c81c2a-7ad0-4c76-868a-978b40d5b74a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Port 154da9f6-d230-479a-88e0-f13ebc655493 binding to destination host cpu-1 is already ACTIVE {{(pid=63088) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1219.329227] env[63088]: DEBUG oslo_concurrency.lockutils [None req-e0c81c2a-7ad0-4c76-868a-978b40d5b74a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquiring lock "refresh_cache-f721d879-55a5-42d1-99c5-871d2e9103ec" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1219.329358] env[63088]: DEBUG oslo_concurrency.lockutils [None req-e0c81c2a-7ad0-4c76-868a-978b40d5b74a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquired lock "refresh_cache-f721d879-55a5-42d1-99c5-871d2e9103ec" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1219.329530] env[63088]: DEBUG nova.network.neutron [None req-e0c81c2a-7ad0-4c76-868a-978b40d5b74a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1219.466550] env[63088]: DEBUG oslo_concurrency.lockutils [req-0e1636e6-36b0-4dbe-a498-f007a880e127 req-7bcbc400-8563-472e-820d-fef7a2a3603e service nova] Releasing lock "refresh_cache-8e1bb264-bd9a-4779-9bbf-849c14c650a9" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1219.676968] env[63088]: DEBUG oslo_vmware.api [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52fa3614-9010-bc98-2665-149353ca8d30, 'name': SearchDatastore_Task, 'duration_secs': 0.007591} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1219.677748] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e7c602d4-ad39-400b-a50a-ac83e0451e2b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.682633] env[63088]: DEBUG oslo_vmware.api [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Waiting for the task: (returnval){ [ 1219.682633] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52a4087e-6aaa-cd88-09ca-a79c68b2beda" [ 1219.682633] env[63088]: _type = "Task" [ 1219.682633] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1219.690135] env[63088]: DEBUG oslo_vmware.api [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52a4087e-6aaa-cd88-09ca-a79c68b2beda, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1220.040877] env[63088]: DEBUG nova.network.neutron [None req-e0c81c2a-7ad0-4c76-868a-978b40d5b74a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Updating instance_info_cache with network_info: [{"id": "154da9f6-d230-479a-88e0-f13ebc655493", "address": "fa:16:3e:47:57:05", "network": {"id": "711d3d9a-86db-4aa1-87b5-481f1812dfa4", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-59376186-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.225", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a7898ddafe0d41038a6ae4277f2c2f48", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e41070eb-3ac1-4ca9-a3d0-fd65893a97de", "external-id": "nsx-vlan-transportzone-596", "segmentation_id": 596, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap154da9f6-d2", "ovs_interfaceid": "154da9f6-d230-479a-88e0-f13ebc655493", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1220.192987] env[63088]: DEBUG oslo_vmware.api [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52a4087e-6aaa-cd88-09ca-a79c68b2beda, 'name': SearchDatastore_Task, 'duration_secs': 0.008663} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1220.193279] env[63088]: DEBUG oslo_concurrency.lockutils [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1220.193548] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] 8e1bb264-bd9a-4779-9bbf-849c14c650a9/8e1bb264-bd9a-4779-9bbf-849c14c650a9.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1220.193813] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-966719f7-cf95-48e0-ae50-6c02402c6cc8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.202376] env[63088]: DEBUG oslo_vmware.api [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Waiting for the task: (returnval){ [ 1220.202376] env[63088]: value = "task-1285611" [ 1220.202376] env[63088]: _type = "Task" [ 1220.202376] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1220.208852] env[63088]: DEBUG oslo_vmware.api [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Task: {'id': task-1285611, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1220.543827] env[63088]: DEBUG oslo_concurrency.lockutils [None req-e0c81c2a-7ad0-4c76-868a-978b40d5b74a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Releasing lock "refresh_cache-f721d879-55a5-42d1-99c5-871d2e9103ec" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1220.710992] env[63088]: DEBUG oslo_vmware.api [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Task: {'id': task-1285611, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.491909} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1220.711290] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] 8e1bb264-bd9a-4779-9bbf-849c14c650a9/8e1bb264-bd9a-4779-9bbf-849c14c650a9.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1220.711508] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] [instance: 8e1bb264-bd9a-4779-9bbf-849c14c650a9] Extending root virtual disk to 1048576 {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1220.711774] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-90982663-b280-49e0-953a-a9db93a9d223 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.717691] env[63088]: DEBUG oslo_vmware.api [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Waiting for the task: (returnval){ [ 1220.717691] env[63088]: value = "task-1285612" [ 1220.717691] env[63088]: _type = "Task" [ 1220.717691] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1220.725159] env[63088]: DEBUG oslo_vmware.api [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Task: {'id': task-1285612, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1221.048057] env[63088]: DEBUG nova.compute.manager [None req-e0c81c2a-7ad0-4c76-868a-978b40d5b74a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=63088) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:897}} [ 1221.227223] env[63088]: DEBUG oslo_vmware.api [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Task: {'id': task-1285612, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.345783} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1221.227507] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] [instance: 8e1bb264-bd9a-4779-9bbf-849c14c650a9] Extended root virtual disk {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1221.228283] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7718540f-4c0b-4132-bb94-810bb2e767bb {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.249442] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] [instance: 8e1bb264-bd9a-4779-9bbf-849c14c650a9] Reconfiguring VM instance instance-0000006b to attach disk [datastore1] 8e1bb264-bd9a-4779-9bbf-849c14c650a9/8e1bb264-bd9a-4779-9bbf-849c14c650a9.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1221.249719] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-58cce172-fa5a-4cfd-993a-71958b44b7a7 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.269248] env[63088]: DEBUG oslo_vmware.api [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Waiting for the task: (returnval){ [ 1221.269248] env[63088]: value = "task-1285613" [ 1221.269248] env[63088]: _type = "Task" [ 1221.269248] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1221.277120] env[63088]: DEBUG oslo_vmware.api [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Task: {'id': task-1285613, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1221.604760] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1221.605197] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Starting heal instance info cache {{(pid=63088) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1221.605197] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Rebuilding the list of instances to heal {{(pid=63088) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 1221.780291] env[63088]: DEBUG oslo_vmware.api [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Task: {'id': task-1285613, 'name': ReconfigVM_Task, 'duration_secs': 0.498488} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1221.780567] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] [instance: 8e1bb264-bd9a-4779-9bbf-849c14c650a9] Reconfigured VM instance instance-0000006b to attach disk [datastore1] 8e1bb264-bd9a-4779-9bbf-849c14c650a9/8e1bb264-bd9a-4779-9bbf-849c14c650a9.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1221.781165] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b90bb3d1-aa00-4d6a-8762-064e00a2f020 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.786829] env[63088]: DEBUG oslo_vmware.api [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Waiting for the task: (returnval){ [ 1221.786829] env[63088]: value = "task-1285614" [ 1221.786829] env[63088]: _type = "Task" [ 1221.786829] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1221.794722] env[63088]: DEBUG oslo_vmware.api [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Task: {'id': task-1285614, 'name': Rename_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1222.108766] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: 8e1bb264-bd9a-4779-9bbf-849c14c650a9] Skipping network cache update for instance because it is Building. {{(pid=63088) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 1222.109035] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Acquiring lock "refresh_cache-f721d879-55a5-42d1-99c5-871d2e9103ec" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1222.109163] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Acquired lock "refresh_cache-f721d879-55a5-42d1-99c5-871d2e9103ec" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1222.109300] env[63088]: DEBUG nova.network.neutron [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Forcefully refreshing network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1222.109446] env[63088]: DEBUG nova.objects.instance [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Lazy-loading 'info_cache' on Instance uuid f721d879-55a5-42d1-99c5-871d2e9103ec {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1222.151412] env[63088]: DEBUG oslo_concurrency.lockutils [None req-e0c81c2a-7ad0-4c76-868a-978b40d5b74a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1222.151688] env[63088]: DEBUG oslo_concurrency.lockutils [None req-e0c81c2a-7ad0-4c76-868a-978b40d5b74a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1222.298641] env[63088]: DEBUG oslo_vmware.api [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Task: {'id': task-1285614, 'name': Rename_Task, 'duration_secs': 0.124072} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1222.298927] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] [instance: 8e1bb264-bd9a-4779-9bbf-849c14c650a9] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1222.299190] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-26f4deca-032e-4581-b33e-622dbe2d4ac8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.305101] env[63088]: DEBUG oslo_vmware.api [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Waiting for the task: (returnval){ [ 1222.305101] env[63088]: value = "task-1285615" [ 1222.305101] env[63088]: _type = "Task" [ 1222.305101] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1222.313676] env[63088]: DEBUG oslo_vmware.api [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Task: {'id': task-1285615, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1222.654622] env[63088]: DEBUG nova.objects.instance [None req-e0c81c2a-7ad0-4c76-868a-978b40d5b74a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lazy-loading 'migration_context' on Instance uuid f721d879-55a5-42d1-99c5-871d2e9103ec {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1222.814062] env[63088]: DEBUG oslo_vmware.api [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Task: {'id': task-1285615, 'name': PowerOnVM_Task, 'duration_secs': 0.451367} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1222.814361] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] [instance: 8e1bb264-bd9a-4779-9bbf-849c14c650a9] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1222.814570] env[63088]: INFO nova.compute.manager [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] [instance: 8e1bb264-bd9a-4779-9bbf-849c14c650a9] Took 6.68 seconds to spawn the instance on the hypervisor. [ 1222.814753] env[63088]: DEBUG nova.compute.manager [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] [instance: 8e1bb264-bd9a-4779-9bbf-849c14c650a9] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1222.815565] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cf18f45-4c18-4f08-b708-2f15949eb476 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.211843] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9286f564-ffba-4bf3-8696-1a9cfe927e1e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.218879] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e6b2038-e264-4a2e-bbae-e3dac412d821 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.249410] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25eed664-c4a3-4e14-9385-b1f25401b6c8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.256371] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ccd8031-ee4f-4969-9984-9b8b9084947d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.269378] env[63088]: DEBUG nova.compute.provider_tree [None req-e0c81c2a-7ad0-4c76-868a-978b40d5b74a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1223.329857] env[63088]: INFO nova.compute.manager [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] [instance: 8e1bb264-bd9a-4779-9bbf-849c14c650a9] Took 11.35 seconds to build instance. [ 1223.771881] env[63088]: DEBUG nova.scheduler.client.report [None req-e0c81c2a-7ad0-4c76-868a-978b40d5b74a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1223.831815] env[63088]: DEBUG oslo_concurrency.lockutils [None req-574ef541-42c4-424e-a9a1-713084c67376 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Lock "8e1bb264-bd9a-4779-9bbf-849c14c650a9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.863s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1223.836500] env[63088]: DEBUG nova.network.neutron [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Updating instance_info_cache with network_info: [{"id": "154da9f6-d230-479a-88e0-f13ebc655493", "address": "fa:16:3e:47:57:05", "network": {"id": "711d3d9a-86db-4aa1-87b5-481f1812dfa4", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-59376186-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.225", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a7898ddafe0d41038a6ae4277f2c2f48", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e41070eb-3ac1-4ca9-a3d0-fd65893a97de", "external-id": "nsx-vlan-transportzone-596", "segmentation_id": 596, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap154da9f6-d2", "ovs_interfaceid": "154da9f6-d230-479a-88e0-f13ebc655493", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1224.339296] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Releasing lock "refresh_cache-f721d879-55a5-42d1-99c5-871d2e9103ec" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1224.339296] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Updated the network info_cache for instance {{(pid=63088) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 1224.339462] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1224.339606] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1224.339805] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager.update_available_resource {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1224.782601] env[63088]: DEBUG oslo_concurrency.lockutils [None req-e0c81c2a-7ad0-4c76-868a-978b40d5b74a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.631s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1224.789570] env[63088]: DEBUG oslo_concurrency.lockutils [None req-1182facc-6230-467f-a66c-f43dbd2fee31 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Acquiring lock "8e1bb264-bd9a-4779-9bbf-849c14c650a9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1224.789797] env[63088]: DEBUG oslo_concurrency.lockutils [None req-1182facc-6230-467f-a66c-f43dbd2fee31 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Lock "8e1bb264-bd9a-4779-9bbf-849c14c650a9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1224.790010] env[63088]: DEBUG oslo_concurrency.lockutils [None req-1182facc-6230-467f-a66c-f43dbd2fee31 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Acquiring lock "8e1bb264-bd9a-4779-9bbf-849c14c650a9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1224.790192] env[63088]: DEBUG oslo_concurrency.lockutils [None req-1182facc-6230-467f-a66c-f43dbd2fee31 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Lock "8e1bb264-bd9a-4779-9bbf-849c14c650a9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1224.790366] env[63088]: DEBUG oslo_concurrency.lockutils [None req-1182facc-6230-467f-a66c-f43dbd2fee31 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Lock "8e1bb264-bd9a-4779-9bbf-849c14c650a9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1224.796135] env[63088]: INFO nova.compute.manager [None req-1182facc-6230-467f-a66c-f43dbd2fee31 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] [instance: 8e1bb264-bd9a-4779-9bbf-849c14c650a9] Terminating instance [ 1224.799260] env[63088]: DEBUG nova.compute.manager [None req-1182facc-6230-467f-a66c-f43dbd2fee31 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] [instance: 8e1bb264-bd9a-4779-9bbf-849c14c650a9] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1224.799644] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-1182facc-6230-467f-a66c-f43dbd2fee31 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] [instance: 8e1bb264-bd9a-4779-9bbf-849c14c650a9] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1224.800544] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f9442b1-c57c-4476-a374-bca9631f25ae {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.808717] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-1182facc-6230-467f-a66c-f43dbd2fee31 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] [instance: 8e1bb264-bd9a-4779-9bbf-849c14c650a9] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1224.808962] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2fd56d22-c0b0-4109-9281-3a94f3ac28a2 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.814883] env[63088]: DEBUG oslo_vmware.api [None req-1182facc-6230-467f-a66c-f43dbd2fee31 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Waiting for the task: (returnval){ [ 1224.814883] env[63088]: value = "task-1285616" [ 1224.814883] env[63088]: _type = "Task" [ 1224.814883] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1224.822750] env[63088]: DEBUG oslo_vmware.api [None req-1182facc-6230-467f-a66c-f43dbd2fee31 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Task: {'id': task-1285616, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1224.842662] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1224.842897] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1224.843079] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1224.843242] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63088) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1224.844120] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f31d11c-e0d4-4e70-ba16-772c38804bb2 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.851481] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30b54f68-41ad-4973-a165-5ecc1b091420 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.866317] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db1c8aa5-2584-423a-9fd4-0424cd5d3d0a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.872881] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ad6dead-e044-4737-a1b1-3f9c947d7507 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.901154] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181250MB free_disk=140GB free_vcpus=48 pci_devices=None {{(pid=63088) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1224.901319] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1224.901501] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1225.324910] env[63088]: DEBUG oslo_vmware.api [None req-1182facc-6230-467f-a66c-f43dbd2fee31 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Task: {'id': task-1285616, 'name': PowerOffVM_Task, 'duration_secs': 0.191078} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1225.325208] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-1182facc-6230-467f-a66c-f43dbd2fee31 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] [instance: 8e1bb264-bd9a-4779-9bbf-849c14c650a9] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1225.325373] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-1182facc-6230-467f-a66c-f43dbd2fee31 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] [instance: 8e1bb264-bd9a-4779-9bbf-849c14c650a9] Unregistering the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1225.325624] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-790d8678-01a7-42f7-8b02-b732689b2d32 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.383970] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-1182facc-6230-467f-a66c-f43dbd2fee31 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] [instance: 8e1bb264-bd9a-4779-9bbf-849c14c650a9] Unregistered the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1225.384278] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-1182facc-6230-467f-a66c-f43dbd2fee31 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] [instance: 8e1bb264-bd9a-4779-9bbf-849c14c650a9] Deleting contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1225.384486] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-1182facc-6230-467f-a66c-f43dbd2fee31 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Deleting the datastore file [datastore1] 8e1bb264-bd9a-4779-9bbf-849c14c650a9 {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1225.384752] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-880d1d7a-ae59-45ae-ada5-e18524ab2f3d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.390784] env[63088]: DEBUG oslo_vmware.api [None req-1182facc-6230-467f-a66c-f43dbd2fee31 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Waiting for the task: (returnval){ [ 1225.390784] env[63088]: value = "task-1285618" [ 1225.390784] env[63088]: _type = "Task" [ 1225.390784] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1225.398728] env[63088]: DEBUG oslo_vmware.api [None req-1182facc-6230-467f-a66c-f43dbd2fee31 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Task: {'id': task-1285618, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1225.900663] env[63088]: DEBUG oslo_vmware.api [None req-1182facc-6230-467f-a66c-f43dbd2fee31 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Task: {'id': task-1285618, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.152764} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1225.901061] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-1182facc-6230-467f-a66c-f43dbd2fee31 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1225.901212] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-1182facc-6230-467f-a66c-f43dbd2fee31 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] [instance: 8e1bb264-bd9a-4779-9bbf-849c14c650a9] Deleted contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1225.901397] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-1182facc-6230-467f-a66c-f43dbd2fee31 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] [instance: 8e1bb264-bd9a-4779-9bbf-849c14c650a9] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1225.901580] env[63088]: INFO nova.compute.manager [None req-1182facc-6230-467f-a66c-f43dbd2fee31 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] [instance: 8e1bb264-bd9a-4779-9bbf-849c14c650a9] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1225.901878] env[63088]: DEBUG oslo.service.loopingcall [None req-1182facc-6230-467f-a66c-f43dbd2fee31 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1225.902116] env[63088]: DEBUG nova.compute.manager [-] [instance: 8e1bb264-bd9a-4779-9bbf-849c14c650a9] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1225.902222] env[63088]: DEBUG nova.network.neutron [-] [instance: 8e1bb264-bd9a-4779-9bbf-849c14c650a9] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1226.125358] env[63088]: DEBUG nova.compute.manager [req-d1ee2f55-4fef-4206-b618-376f3f091aa6 req-862d120a-60ba-477c-9e75-21de9fb8eae0 service nova] [instance: 8e1bb264-bd9a-4779-9bbf-849c14c650a9] Received event network-vif-deleted-de4db792-3bff-4ed9-875e-8bfaeed5d0ab {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1226.125415] env[63088]: INFO nova.compute.manager [req-d1ee2f55-4fef-4206-b618-376f3f091aa6 req-862d120a-60ba-477c-9e75-21de9fb8eae0 service nova] [instance: 8e1bb264-bd9a-4779-9bbf-849c14c650a9] Neutron deleted interface de4db792-3bff-4ed9-875e-8bfaeed5d0ab; detaching it from the instance and deleting it from the info cache [ 1226.125604] env[63088]: DEBUG nova.network.neutron [req-d1ee2f55-4fef-4206-b618-376f3f091aa6 req-862d120a-60ba-477c-9e75-21de9fb8eae0 service nova] [instance: 8e1bb264-bd9a-4779-9bbf-849c14c650a9] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1226.321968] env[63088]: INFO nova.compute.manager [None req-e0c81c2a-7ad0-4c76-868a-978b40d5b74a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Swapping old allocation on dict_keys(['6eae54a9-8831-40eb-bf54-4bc60d346b02']) held by migration 1047a6e9-8d90-4ad4-b322-8fe7c2def26c for instance [ 1226.342318] env[63088]: DEBUG nova.scheduler.client.report [None req-e0c81c2a-7ad0-4c76-868a-978b40d5b74a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Overwriting current allocation {'allocations': {'6eae54a9-8831-40eb-bf54-4bc60d346b02': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 141}}, 'project_id': 'a7898ddafe0d41038a6ae4277f2c2f48', 'user_id': '6ffb8b92c6d649c4b39d8df1c1d0c36c', 'consumer_generation': 1} on consumer f721d879-55a5-42d1-99c5-871d2e9103ec {{(pid=63088) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 1226.415925] env[63088]: DEBUG oslo_concurrency.lockutils [None req-e0c81c2a-7ad0-4c76-868a-978b40d5b74a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquiring lock "refresh_cache-f721d879-55a5-42d1-99c5-871d2e9103ec" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1226.416186] env[63088]: DEBUG oslo_concurrency.lockutils [None req-e0c81c2a-7ad0-4c76-868a-978b40d5b74a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquired lock "refresh_cache-f721d879-55a5-42d1-99c5-871d2e9103ec" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1226.416428] env[63088]: DEBUG nova.network.neutron [None req-e0c81c2a-7ad0-4c76-868a-978b40d5b74a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1226.427745] env[63088]: INFO nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance 1047a6e9-8d90-4ad4-b322-8fe7c2def26c has allocations against this compute host but is not found in the database. [ 1226.428240] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance f721d879-55a5-42d1-99c5-871d2e9103ec actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1226.428406] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance 8e1bb264-bd9a-4779-9bbf-849c14c650a9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1226.428578] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Total usable vcpus: 48, total allocated vcpus: 2 {{(pid=63088) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1226.428715] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=960MB phys_disk=200GB used_disk=2GB total_vcpus=48 used_vcpus=2 pci_stats=[] {{(pid=63088) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1226.463325] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a34ed804-1c5e-459f-b0a1-c1b0e2801979 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.470804] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-017c847d-d627-4656-8c1c-6cec9145f70c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.499810] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b42c2b1a-d47d-42aa-b0d1-cf9e99dd5d58 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.506424] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d55fabd-48e2-40a5-a00a-80dbf947a3ee {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.519073] env[63088]: DEBUG nova.compute.provider_tree [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1226.604156] env[63088]: DEBUG nova.network.neutron [-] [instance: 8e1bb264-bd9a-4779-9bbf-849c14c650a9] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1226.628410] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8635107d-f221-4169-a691-2774568fa23f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.637837] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2780825f-faec-440e-9afe-e9d9bfe3aa20 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.660401] env[63088]: DEBUG nova.compute.manager [req-d1ee2f55-4fef-4206-b618-376f3f091aa6 req-862d120a-60ba-477c-9e75-21de9fb8eae0 service nova] [instance: 8e1bb264-bd9a-4779-9bbf-849c14c650a9] Detach interface failed, port_id=de4db792-3bff-4ed9-875e-8bfaeed5d0ab, reason: Instance 8e1bb264-bd9a-4779-9bbf-849c14c650a9 could not be found. {{(pid=63088) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1227.021693] env[63088]: DEBUG nova.scheduler.client.report [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1227.107104] env[63088]: INFO nova.compute.manager [-] [instance: 8e1bb264-bd9a-4779-9bbf-849c14c650a9] Took 1.20 seconds to deallocate network for instance. [ 1227.123121] env[63088]: DEBUG nova.network.neutron [None req-e0c81c2a-7ad0-4c76-868a-978b40d5b74a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Updating instance_info_cache with network_info: [{"id": "154da9f6-d230-479a-88e0-f13ebc655493", "address": "fa:16:3e:47:57:05", "network": {"id": "711d3d9a-86db-4aa1-87b5-481f1812dfa4", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-59376186-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.225", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a7898ddafe0d41038a6ae4277f2c2f48", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e41070eb-3ac1-4ca9-a3d0-fd65893a97de", "external-id": "nsx-vlan-transportzone-596", "segmentation_id": 596, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap154da9f6-d2", "ovs_interfaceid": "154da9f6-d230-479a-88e0-f13ebc655493", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1227.526819] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63088) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1227.527048] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.625s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1227.613434] env[63088]: DEBUG oslo_concurrency.lockutils [None req-1182facc-6230-467f-a66c-f43dbd2fee31 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1227.613724] env[63088]: DEBUG oslo_concurrency.lockutils [None req-1182facc-6230-467f-a66c-f43dbd2fee31 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1227.613956] env[63088]: DEBUG nova.objects.instance [None req-1182facc-6230-467f-a66c-f43dbd2fee31 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Lazy-loading 'resources' on Instance uuid 8e1bb264-bd9a-4779-9bbf-849c14c650a9 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1227.625429] env[63088]: DEBUG oslo_concurrency.lockutils [None req-e0c81c2a-7ad0-4c76-868a-978b40d5b74a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Releasing lock "refresh_cache-f721d879-55a5-42d1-99c5-871d2e9103ec" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1227.626073] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c2934c9-309f-4fe5-8ded-6c93ba270e93 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.633900] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49283f43-d4ba-4ff6-bb3a-349ae6a34f3a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.160136] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d60d0580-5b3a-4411-bce8-7b0b5ca20aec {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.167674] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62ba0572-c4e1-46b8-921b-72188d7b39b6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.201317] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eaca5b9c-d8ff-4b2e-9539-58ae045f69b1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.208240] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e68c91c1-2286-475f-bbfd-498d9232c9f9 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.221174] env[63088]: DEBUG nova.compute.provider_tree [None req-1182facc-6230-467f-a66c-f43dbd2fee31 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1228.523115] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1228.523115] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1228.724323] env[63088]: DEBUG nova.scheduler.client.report [None req-1182facc-6230-467f-a66c-f43dbd2fee31 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1228.727904] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0c81c2a-7ad0-4c76-868a-978b40d5b74a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1228.728391] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8c3eb1a3-a5d1-452f-8079-20ab7a091059 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.735482] env[63088]: DEBUG oslo_vmware.api [None req-e0c81c2a-7ad0-4c76-868a-978b40d5b74a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1228.735482] env[63088]: value = "task-1285619" [ 1228.735482] env[63088]: _type = "Task" [ 1228.735482] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1228.744087] env[63088]: DEBUG oslo_vmware.api [None req-e0c81c2a-7ad0-4c76-868a-978b40d5b74a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285619, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1229.229352] env[63088]: DEBUG oslo_concurrency.lockutils [None req-1182facc-6230-467f-a66c-f43dbd2fee31 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.615s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1229.245056] env[63088]: DEBUG oslo_vmware.api [None req-e0c81c2a-7ad0-4c76-868a-978b40d5b74a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285619, 'name': PowerOffVM_Task, 'duration_secs': 0.174742} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1229.245370] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0c81c2a-7ad0-4c76-868a-978b40d5b74a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1229.246113] env[63088]: DEBUG nova.virt.hardware [None req-e0c81c2a-7ad0-4c76-868a-978b40d5b74a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1229.246422] env[63088]: DEBUG nova.virt.hardware [None req-e0c81c2a-7ad0-4c76-868a-978b40d5b74a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1229.246633] env[63088]: DEBUG nova.virt.hardware [None req-e0c81c2a-7ad0-4c76-868a-978b40d5b74a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1229.246861] env[63088]: DEBUG nova.virt.hardware [None req-e0c81c2a-7ad0-4c76-868a-978b40d5b74a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1229.247579] env[63088]: DEBUG nova.virt.hardware [None req-e0c81c2a-7ad0-4c76-868a-978b40d5b74a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1229.247814] env[63088]: DEBUG nova.virt.hardware [None req-e0c81c2a-7ad0-4c76-868a-978b40d5b74a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1229.248086] env[63088]: DEBUG nova.virt.hardware [None req-e0c81c2a-7ad0-4c76-868a-978b40d5b74a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1229.248292] env[63088]: DEBUG nova.virt.hardware [None req-e0c81c2a-7ad0-4c76-868a-978b40d5b74a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1229.248505] env[63088]: DEBUG nova.virt.hardware [None req-e0c81c2a-7ad0-4c76-868a-978b40d5b74a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1229.248703] env[63088]: DEBUG nova.virt.hardware [None req-e0c81c2a-7ad0-4c76-868a-978b40d5b74a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1229.248906] env[63088]: DEBUG nova.virt.hardware [None req-e0c81c2a-7ad0-4c76-868a-978b40d5b74a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1229.255046] env[63088]: INFO nova.scheduler.client.report [None req-1182facc-6230-467f-a66c-f43dbd2fee31 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Deleted allocations for instance 8e1bb264-bd9a-4779-9bbf-849c14c650a9 [ 1229.256049] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2b75d195-e0fc-4c79-9736-fec8e94d80d8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.273445] env[63088]: DEBUG oslo_vmware.api [None req-e0c81c2a-7ad0-4c76-868a-978b40d5b74a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1229.273445] env[63088]: value = "task-1285620" [ 1229.273445] env[63088]: _type = "Task" [ 1229.273445] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1229.281921] env[63088]: DEBUG oslo_vmware.api [None req-e0c81c2a-7ad0-4c76-868a-978b40d5b74a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285620, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1229.772653] env[63088]: DEBUG oslo_concurrency.lockutils [None req-1182facc-6230-467f-a66c-f43dbd2fee31 tempest-ServerTagsTestJSON-1512523977 tempest-ServerTagsTestJSON-1512523977-project-member] Lock "8e1bb264-bd9a-4779-9bbf-849c14c650a9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.983s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1229.781832] env[63088]: DEBUG oslo_vmware.api [None req-e0c81c2a-7ad0-4c76-868a-978b40d5b74a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285620, 'name': ReconfigVM_Task, 'duration_secs': 0.139308} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1229.784045] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c85ca38-174e-404c-9f5e-c0d3028aa2f2 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.803732] env[63088]: DEBUG nova.virt.hardware [None req-e0c81c2a-7ad0-4c76-868a-978b40d5b74a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1229.804187] env[63088]: DEBUG nova.virt.hardware [None req-e0c81c2a-7ad0-4c76-868a-978b40d5b74a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1229.804187] env[63088]: DEBUG nova.virt.hardware [None req-e0c81c2a-7ad0-4c76-868a-978b40d5b74a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1229.804332] env[63088]: DEBUG nova.virt.hardware [None req-e0c81c2a-7ad0-4c76-868a-978b40d5b74a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1229.804497] env[63088]: DEBUG nova.virt.hardware [None req-e0c81c2a-7ad0-4c76-868a-978b40d5b74a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1229.804652] env[63088]: DEBUG nova.virt.hardware [None req-e0c81c2a-7ad0-4c76-868a-978b40d5b74a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1229.804854] env[63088]: DEBUG nova.virt.hardware [None req-e0c81c2a-7ad0-4c76-868a-978b40d5b74a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1229.805021] env[63088]: DEBUG nova.virt.hardware [None req-e0c81c2a-7ad0-4c76-868a-978b40d5b74a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1229.805213] env[63088]: DEBUG nova.virt.hardware [None req-e0c81c2a-7ad0-4c76-868a-978b40d5b74a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1229.805390] env[63088]: DEBUG nova.virt.hardware [None req-e0c81c2a-7ad0-4c76-868a-978b40d5b74a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1229.805576] env[63088]: DEBUG nova.virt.hardware [None req-e0c81c2a-7ad0-4c76-868a-978b40d5b74a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1229.806895] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d7e45fd6-d462-480e-be33-f000734bbe48 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.812446] env[63088]: DEBUG oslo_vmware.api [None req-e0c81c2a-7ad0-4c76-868a-978b40d5b74a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1229.812446] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]5234f8a3-fadf-7af0-3020-9e5bdbcb3d0a" [ 1229.812446] env[63088]: _type = "Task" [ 1229.812446] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1229.820249] env[63088]: DEBUG oslo_vmware.api [None req-e0c81c2a-7ad0-4c76-868a-978b40d5b74a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]5234f8a3-fadf-7af0-3020-9e5bdbcb3d0a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1230.322512] env[63088]: DEBUG oslo_vmware.api [None req-e0c81c2a-7ad0-4c76-868a-978b40d5b74a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]5234f8a3-fadf-7af0-3020-9e5bdbcb3d0a, 'name': SearchDatastore_Task, 'duration_secs': 0.008888} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1230.327967] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-e0c81c2a-7ad0-4c76-868a-978b40d5b74a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Reconfiguring VM instance instance-0000006a to detach disk 2000 {{(pid=63088) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1230.328581] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b16fd2bd-6021-4982-a1d3-2bc06b1b155d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.346484] env[63088]: DEBUG oslo_vmware.api [None req-e0c81c2a-7ad0-4c76-868a-978b40d5b74a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1230.346484] env[63088]: value = "task-1285621" [ 1230.346484] env[63088]: _type = "Task" [ 1230.346484] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1230.354598] env[63088]: DEBUG oslo_vmware.api [None req-e0c81c2a-7ad0-4c76-868a-978b40d5b74a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285621, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1230.856737] env[63088]: DEBUG oslo_vmware.api [None req-e0c81c2a-7ad0-4c76-868a-978b40d5b74a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285621, 'name': ReconfigVM_Task, 'duration_secs': 0.193368} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1230.857030] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-e0c81c2a-7ad0-4c76-868a-978b40d5b74a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Reconfigured VM instance instance-0000006a to detach disk 2000 {{(pid=63088) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1230.858032] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb14a9b3-a0a9-4cc0-a0a2-cc2a6d6612eb {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.883334] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-e0c81c2a-7ad0-4c76-868a-978b40d5b74a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Reconfiguring VM instance instance-0000006a to attach disk [datastore1] f721d879-55a5-42d1-99c5-871d2e9103ec/f721d879-55a5-42d1-99c5-871d2e9103ec.vmdk or device None with type thin {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1230.883677] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cefae9d5-bba7-47e4-bc6c-1b440e595099 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.902935] env[63088]: DEBUG oslo_vmware.api [None req-e0c81c2a-7ad0-4c76-868a-978b40d5b74a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1230.902935] env[63088]: value = "task-1285622" [ 1230.902935] env[63088]: _type = "Task" [ 1230.902935] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1230.911271] env[63088]: DEBUG oslo_vmware.api [None req-e0c81c2a-7ad0-4c76-868a-978b40d5b74a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285622, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1231.413109] env[63088]: DEBUG oslo_vmware.api [None req-e0c81c2a-7ad0-4c76-868a-978b40d5b74a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285622, 'name': ReconfigVM_Task, 'duration_secs': 0.306682} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1231.413405] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-e0c81c2a-7ad0-4c76-868a-978b40d5b74a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Reconfigured VM instance instance-0000006a to attach disk [datastore1] f721d879-55a5-42d1-99c5-871d2e9103ec/f721d879-55a5-42d1-99c5-871d2e9103ec.vmdk or device None with type thin {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1231.414220] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e04af8a-7ad3-4ae6-8736-be0f8805ac47 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.434124] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-864f5c2a-28e6-4d64-8314-10f795203ccc {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.454835] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d85b3fa0-d733-4b74-8661-049e67a1f11c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.474424] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93d2af10-e811-41ee-9dc4-7ccccecfcfb1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.480558] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0c81c2a-7ad0-4c76-868a-978b40d5b74a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1231.480780] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ed87e1db-ca8d-45c5-ba61-830dccfea121 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.486513] env[63088]: DEBUG oslo_vmware.api [None req-e0c81c2a-7ad0-4c76-868a-978b40d5b74a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1231.486513] env[63088]: value = "task-1285623" [ 1231.486513] env[63088]: _type = "Task" [ 1231.486513] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1231.493490] env[63088]: DEBUG oslo_vmware.api [None req-e0c81c2a-7ad0-4c76-868a-978b40d5b74a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285623, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1231.998742] env[63088]: DEBUG oslo_vmware.api [None req-e0c81c2a-7ad0-4c76-868a-978b40d5b74a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285623, 'name': PowerOnVM_Task, 'duration_secs': 0.365665} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1231.999074] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0c81c2a-7ad0-4c76-868a-978b40d5b74a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1233.050515] env[63088]: INFO nova.compute.manager [None req-e0c81c2a-7ad0-4c76-868a-978b40d5b74a tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Updating instance to original state: 'active' [ 1234.182330] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6af9695f-6439-4cfd-bd87-9708d2504561 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquiring lock "f721d879-55a5-42d1-99c5-871d2e9103ec" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1234.182330] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6af9695f-6439-4cfd-bd87-9708d2504561 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lock "f721d879-55a5-42d1-99c5-871d2e9103ec" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1234.182330] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6af9695f-6439-4cfd-bd87-9708d2504561 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquiring lock "f721d879-55a5-42d1-99c5-871d2e9103ec-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1234.182330] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6af9695f-6439-4cfd-bd87-9708d2504561 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lock "f721d879-55a5-42d1-99c5-871d2e9103ec-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1234.182330] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6af9695f-6439-4cfd-bd87-9708d2504561 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lock "f721d879-55a5-42d1-99c5-871d2e9103ec-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1234.186185] env[63088]: INFO nova.compute.manager [None req-6af9695f-6439-4cfd-bd87-9708d2504561 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Terminating instance [ 1234.188464] env[63088]: DEBUG nova.compute.manager [None req-6af9695f-6439-4cfd-bd87-9708d2504561 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1234.188685] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-6af9695f-6439-4cfd-bd87-9708d2504561 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1234.188951] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ba1d0bd9-1368-4463-897c-4222a0757d36 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.197566] env[63088]: DEBUG oslo_vmware.api [None req-6af9695f-6439-4cfd-bd87-9708d2504561 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1234.197566] env[63088]: value = "task-1285624" [ 1234.197566] env[63088]: _type = "Task" [ 1234.197566] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1234.207181] env[63088]: DEBUG oslo_vmware.api [None req-6af9695f-6439-4cfd-bd87-9708d2504561 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285624, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1234.707691] env[63088]: DEBUG oslo_vmware.api [None req-6af9695f-6439-4cfd-bd87-9708d2504561 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285624, 'name': PowerOffVM_Task, 'duration_secs': 0.174265} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1234.708022] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-6af9695f-6439-4cfd-bd87-9708d2504561 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1234.708372] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-6af9695f-6439-4cfd-bd87-9708d2504561 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Volume detach. Driver type: vmdk {{(pid=63088) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1234.708489] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-6af9695f-6439-4cfd-bd87-9708d2504561 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-275971', 'volume_id': '713f8e10-f238-4677-942d-cd09feaf27c7', 'name': 'volume-713f8e10-f238-4677-942d-cd09feaf27c7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': 'f721d879-55a5-42d1-99c5-871d2e9103ec', 'attached_at': '2024-10-15T12:25:37.000000', 'detached_at': '', 'volume_id': '713f8e10-f238-4677-942d-cd09feaf27c7', 'serial': '713f8e10-f238-4677-942d-cd09feaf27c7'} {{(pid=63088) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1234.709297] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d6da26e-2f12-4cfb-b988-5acab634eace {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.732045] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0d65cd1-55d6-41d7-89b1-6bc59361801d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.738465] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9668c9ef-47a1-4489-82c9-d55325be67ef {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.757886] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34a3b17c-cbcb-4753-8f89-f24fede396b2 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.771815] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-6af9695f-6439-4cfd-bd87-9708d2504561 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] The volume has not been displaced from its original location: [datastore2] volume-713f8e10-f238-4677-942d-cd09feaf27c7/volume-713f8e10-f238-4677-942d-cd09feaf27c7.vmdk. No consolidation needed. {{(pid=63088) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1234.777064] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-6af9695f-6439-4cfd-bd87-9708d2504561 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Reconfiguring VM instance instance-0000006a to detach disk 2001 {{(pid=63088) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1234.777343] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a8050d5c-b374-4802-9d1f-7384c65e550b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.795625] env[63088]: DEBUG oslo_vmware.api [None req-6af9695f-6439-4cfd-bd87-9708d2504561 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1234.795625] env[63088]: value = "task-1285625" [ 1234.795625] env[63088]: _type = "Task" [ 1234.795625] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1234.802392] env[63088]: DEBUG oslo_vmware.api [None req-6af9695f-6439-4cfd-bd87-9708d2504561 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285625, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1235.145056] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Acquiring lock "918eaae2-f821-402a-a0c2-2c140e3df4d7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1235.145056] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Lock "918eaae2-f821-402a-a0c2-2c140e3df4d7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1235.305674] env[63088]: DEBUG oslo_vmware.api [None req-6af9695f-6439-4cfd-bd87-9708d2504561 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285625, 'name': ReconfigVM_Task, 'duration_secs': 0.187265} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1235.306019] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-6af9695f-6439-4cfd-bd87-9708d2504561 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Reconfigured VM instance instance-0000006a to detach disk 2001 {{(pid=63088) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1235.310633] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-63aa462e-3e82-45ef-a03a-34eaf127b17e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.325445] env[63088]: DEBUG oslo_vmware.api [None req-6af9695f-6439-4cfd-bd87-9708d2504561 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1235.325445] env[63088]: value = "task-1285626" [ 1235.325445] env[63088]: _type = "Task" [ 1235.325445] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1235.332810] env[63088]: DEBUG oslo_vmware.api [None req-6af9695f-6439-4cfd-bd87-9708d2504561 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285626, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1235.646714] env[63088]: DEBUG nova.compute.manager [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1235.834839] env[63088]: DEBUG oslo_vmware.api [None req-6af9695f-6439-4cfd-bd87-9708d2504561 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285626, 'name': ReconfigVM_Task, 'duration_secs': 0.132114} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1235.835166] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-6af9695f-6439-4cfd-bd87-9708d2504561 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-275971', 'volume_id': '713f8e10-f238-4677-942d-cd09feaf27c7', 'name': 'volume-713f8e10-f238-4677-942d-cd09feaf27c7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': 'f721d879-55a5-42d1-99c5-871d2e9103ec', 'attached_at': '2024-10-15T12:25:37.000000', 'detached_at': '', 'volume_id': '713f8e10-f238-4677-942d-cd09feaf27c7', 'serial': '713f8e10-f238-4677-942d-cd09feaf27c7'} {{(pid=63088) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1235.835453] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-6af9695f-6439-4cfd-bd87-9708d2504561 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1235.836222] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8200b86-278b-427b-aa24-fc99d3a6aa05 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.842762] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-6af9695f-6439-4cfd-bd87-9708d2504561 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Unregistering the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1235.842996] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-894f8526-860e-4bcb-9ba5-8ca0436f303e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.905204] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-6af9695f-6439-4cfd-bd87-9708d2504561 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Unregistered the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1235.905420] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-6af9695f-6439-4cfd-bd87-9708d2504561 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Deleting contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1235.905607] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-6af9695f-6439-4cfd-bd87-9708d2504561 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Deleting the datastore file [datastore1] f721d879-55a5-42d1-99c5-871d2e9103ec {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1235.905864] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8fc602af-6e88-41eb-b097-ef84dad7c93f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.911504] env[63088]: DEBUG oslo_vmware.api [None req-6af9695f-6439-4cfd-bd87-9708d2504561 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1235.911504] env[63088]: value = "task-1285628" [ 1235.911504] env[63088]: _type = "Task" [ 1235.911504] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1235.919009] env[63088]: DEBUG oslo_vmware.api [None req-6af9695f-6439-4cfd-bd87-9708d2504561 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285628, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1236.169039] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1236.169217] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1236.170787] env[63088]: INFO nova.compute.claims [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1236.422014] env[63088]: DEBUG oslo_vmware.api [None req-6af9695f-6439-4cfd-bd87-9708d2504561 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285628, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.143256} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1236.422454] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-6af9695f-6439-4cfd-bd87-9708d2504561 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1236.422454] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-6af9695f-6439-4cfd-bd87-9708d2504561 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Deleted contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1236.422588] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-6af9695f-6439-4cfd-bd87-9708d2504561 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1236.422768] env[63088]: INFO nova.compute.manager [None req-6af9695f-6439-4cfd-bd87-9708d2504561 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Took 2.23 seconds to destroy the instance on the hypervisor. [ 1236.423024] env[63088]: DEBUG oslo.service.loopingcall [None req-6af9695f-6439-4cfd-bd87-9708d2504561 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1236.423253] env[63088]: DEBUG nova.compute.manager [-] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1236.423348] env[63088]: DEBUG nova.network.neutron [-] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1236.859070] env[63088]: DEBUG nova.compute.manager [req-0430c129-d596-4488-aa62-d88d10d06e50 req-44bfe4af-a3a8-4d48-b961-4e4ebfa50e78 service nova] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Received event network-vif-deleted-154da9f6-d230-479a-88e0-f13ebc655493 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1236.859211] env[63088]: INFO nova.compute.manager [req-0430c129-d596-4488-aa62-d88d10d06e50 req-44bfe4af-a3a8-4d48-b961-4e4ebfa50e78 service nova] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Neutron deleted interface 154da9f6-d230-479a-88e0-f13ebc655493; detaching it from the instance and deleting it from the info cache [ 1236.859370] env[63088]: DEBUG nova.network.neutron [req-0430c129-d596-4488-aa62-d88d10d06e50 req-44bfe4af-a3a8-4d48-b961-4e4ebfa50e78 service nova] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1237.221818] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b28ac89-8486-47e9-a6d6-72d2564bd150 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.229266] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c30c3be-90bd-4c47-8860-3746308817f3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.259958] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f90b21a5-704f-4d68-91cb-33e71c493716 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.267106] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2e60869-e7eb-42b4-a050-f671acd9d4de {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.279701] env[63088]: DEBUG nova.compute.provider_tree [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1237.336350] env[63088]: DEBUG nova.network.neutron [-] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1237.361996] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-dc8d242f-8b5a-47ba-a1f3-ab7532fdc415 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.372541] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a193949-29b4-457d-8e97-e60b8a696b6f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.392971] env[63088]: DEBUG nova.compute.manager [req-0430c129-d596-4488-aa62-d88d10d06e50 req-44bfe4af-a3a8-4d48-b961-4e4ebfa50e78 service nova] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Detach interface failed, port_id=154da9f6-d230-479a-88e0-f13ebc655493, reason: Instance f721d879-55a5-42d1-99c5-871d2e9103ec could not be found. {{(pid=63088) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1237.782753] env[63088]: DEBUG nova.scheduler.client.report [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1237.838623] env[63088]: INFO nova.compute.manager [-] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Took 1.42 seconds to deallocate network for instance. [ 1238.288109] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.118s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1238.288315] env[63088]: DEBUG nova.compute.manager [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1238.381230] env[63088]: INFO nova.compute.manager [None req-6af9695f-6439-4cfd-bd87-9708d2504561 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Took 0.54 seconds to detach 1 volumes for instance. [ 1238.793387] env[63088]: DEBUG nova.compute.utils [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1238.794872] env[63088]: DEBUG nova.compute.manager [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1238.795065] env[63088]: DEBUG nova.network.neutron [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1238.834650] env[63088]: DEBUG nova.policy [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '11a98d4e5ef3402b99bea85a0bb479e7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e4d8e091fab04519a6c430290830998c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 1238.887788] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6af9695f-6439-4cfd-bd87-9708d2504561 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1238.888069] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6af9695f-6439-4cfd-bd87-9708d2504561 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1238.888296] env[63088]: DEBUG nova.objects.instance [None req-6af9695f-6439-4cfd-bd87-9708d2504561 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lazy-loading 'resources' on Instance uuid f721d879-55a5-42d1-99c5-871d2e9103ec {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1239.106898] env[63088]: DEBUG nova.network.neutron [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] Successfully created port: c1808fff-b8ad-486e-bd90-0f23d115e398 {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1239.298773] env[63088]: DEBUG nova.compute.manager [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1239.435736] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ed28c4a-6dd9-4020-b10a-8c975b8b4cba {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.442860] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-070e1ecb-8176-4155-80a7-05e079d4f5ac {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.471655] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d034de23-f27c-45b6-8775-9000a3b351de {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.478465] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f870a49c-5b90-4a22-8b63-f1635dd07c4e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.490869] env[63088]: DEBUG nova.compute.provider_tree [None req-6af9695f-6439-4cfd-bd87-9708d2504561 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1239.994313] env[63088]: DEBUG nova.scheduler.client.report [None req-6af9695f-6439-4cfd-bd87-9708d2504561 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1240.308287] env[63088]: DEBUG nova.compute.manager [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1240.333192] env[63088]: DEBUG nova.virt.hardware [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1240.333460] env[63088]: DEBUG nova.virt.hardware [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1240.333619] env[63088]: DEBUG nova.virt.hardware [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1240.333806] env[63088]: DEBUG nova.virt.hardware [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1240.333958] env[63088]: DEBUG nova.virt.hardware [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1240.334689] env[63088]: DEBUG nova.virt.hardware [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1240.334997] env[63088]: DEBUG nova.virt.hardware [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1240.335227] env[63088]: DEBUG nova.virt.hardware [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1240.335443] env[63088]: DEBUG nova.virt.hardware [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1240.335625] env[63088]: DEBUG nova.virt.hardware [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1240.335806] env[63088]: DEBUG nova.virt.hardware [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1240.336663] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08948e97-d216-4e36-aea7-914c161c6d68 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.345282] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecde3951-cdbc-4c07-b409-f8a691559cf1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.496083] env[63088]: DEBUG nova.compute.manager [req-9e93c8c2-b5c0-4335-8032-122c59505786 req-45be94bf-8990-4ff2-b70b-045f43045922 service nova] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] Received event network-vif-plugged-c1808fff-b8ad-486e-bd90-0f23d115e398 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1240.496310] env[63088]: DEBUG oslo_concurrency.lockutils [req-9e93c8c2-b5c0-4335-8032-122c59505786 req-45be94bf-8990-4ff2-b70b-045f43045922 service nova] Acquiring lock "918eaae2-f821-402a-a0c2-2c140e3df4d7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1240.496488] env[63088]: DEBUG oslo_concurrency.lockutils [req-9e93c8c2-b5c0-4335-8032-122c59505786 req-45be94bf-8990-4ff2-b70b-045f43045922 service nova] Lock "918eaae2-f821-402a-a0c2-2c140e3df4d7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1240.496678] env[63088]: DEBUG oslo_concurrency.lockutils [req-9e93c8c2-b5c0-4335-8032-122c59505786 req-45be94bf-8990-4ff2-b70b-045f43045922 service nova] Lock "918eaae2-f821-402a-a0c2-2c140e3df4d7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1240.496858] env[63088]: DEBUG nova.compute.manager [req-9e93c8c2-b5c0-4335-8032-122c59505786 req-45be94bf-8990-4ff2-b70b-045f43045922 service nova] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] No waiting events found dispatching network-vif-plugged-c1808fff-b8ad-486e-bd90-0f23d115e398 {{(pid=63088) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1240.497034] env[63088]: WARNING nova.compute.manager [req-9e93c8c2-b5c0-4335-8032-122c59505786 req-45be94bf-8990-4ff2-b70b-045f43045922 service nova] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] Received unexpected event network-vif-plugged-c1808fff-b8ad-486e-bd90-0f23d115e398 for instance with vm_state building and task_state spawning. [ 1240.499052] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6af9695f-6439-4cfd-bd87-9708d2504561 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.611s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1240.516249] env[63088]: INFO nova.scheduler.client.report [None req-6af9695f-6439-4cfd-bd87-9708d2504561 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Deleted allocations for instance f721d879-55a5-42d1-99c5-871d2e9103ec [ 1240.597348] env[63088]: DEBUG nova.network.neutron [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] Successfully updated port: c1808fff-b8ad-486e-bd90-0f23d115e398 {{(pid=63088) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1241.025052] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6af9695f-6439-4cfd-bd87-9708d2504561 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lock "f721d879-55a5-42d1-99c5-871d2e9103ec" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.843s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1241.099914] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Acquiring lock "refresh_cache-918eaae2-f821-402a-a0c2-2c140e3df4d7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1241.100278] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Acquired lock "refresh_cache-918eaae2-f821-402a-a0c2-2c140e3df4d7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1241.100278] env[63088]: DEBUG nova.network.neutron [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1241.730749] env[63088]: DEBUG nova.network.neutron [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1241.920909] env[63088]: DEBUG nova.network.neutron [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] Updating instance_info_cache with network_info: [{"id": "c1808fff-b8ad-486e-bd90-0f23d115e398", "address": "fa:16:3e:15:72:9f", "network": {"id": "ab40c372-54e5-49d6-a02a-279901071b67", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1152691158-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e4d8e091fab04519a6c430290830998c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc1808fff-b8", "ovs_interfaceid": "c1808fff-b8ad-486e-bd90-0f23d115e398", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1242.423425] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Releasing lock "refresh_cache-918eaae2-f821-402a-a0c2-2c140e3df4d7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1242.423787] env[63088]: DEBUG nova.compute.manager [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] Instance network_info: |[{"id": "c1808fff-b8ad-486e-bd90-0f23d115e398", "address": "fa:16:3e:15:72:9f", "network": {"id": "ab40c372-54e5-49d6-a02a-279901071b67", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1152691158-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e4d8e091fab04519a6c430290830998c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc1808fff-b8", "ovs_interfaceid": "c1808fff-b8ad-486e-bd90-0f23d115e398", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1242.424175] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:15:72:9f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '418ddd3d-5f64-407e-8e0c-c8b81639bee9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c1808fff-b8ad-486e-bd90-0f23d115e398', 'vif_model': 'vmxnet3'}] {{(pid=63088) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1242.431545] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Creating folder: Project (e4d8e091fab04519a6c430290830998c). Parent ref: group-v275816. {{(pid=63088) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1242.431825] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cbbc43f2-87bd-4258-998f-6bc4a07dbd50 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.442859] env[63088]: INFO nova.virt.vmwareapi.vm_util [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Created folder: Project (e4d8e091fab04519a6c430290830998c) in parent group-v275816. [ 1242.443054] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Creating folder: Instances. Parent ref: group-v275975. {{(pid=63088) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1242.443279] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3a827e11-0a36-46dc-82d7-bcf068ba76c1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.451836] env[63088]: INFO nova.virt.vmwareapi.vm_util [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Created folder: Instances in parent group-v275975. [ 1242.452065] env[63088]: DEBUG oslo.service.loopingcall [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1242.452250] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] Creating VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1242.452437] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6520799c-32b3-42e9-b3c9-7ecb3ea7ab7a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.470516] env[63088]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1242.470516] env[63088]: value = "task-1285631" [ 1242.470516] env[63088]: _type = "Task" [ 1242.470516] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1242.477463] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285631, 'name': CreateVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1242.521928] env[63088]: DEBUG nova.compute.manager [req-b4c5f0a8-30d7-4dae-9c6b-ab2db7f66603 req-013af72a-9d3f-445c-b00d-81f8d93f24cc service nova] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] Received event network-changed-c1808fff-b8ad-486e-bd90-0f23d115e398 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1242.522188] env[63088]: DEBUG nova.compute.manager [req-b4c5f0a8-30d7-4dae-9c6b-ab2db7f66603 req-013af72a-9d3f-445c-b00d-81f8d93f24cc service nova] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] Refreshing instance network info cache due to event network-changed-c1808fff-b8ad-486e-bd90-0f23d115e398. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1242.522501] env[63088]: DEBUG oslo_concurrency.lockutils [req-b4c5f0a8-30d7-4dae-9c6b-ab2db7f66603 req-013af72a-9d3f-445c-b00d-81f8d93f24cc service nova] Acquiring lock "refresh_cache-918eaae2-f821-402a-a0c2-2c140e3df4d7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1242.522692] env[63088]: DEBUG oslo_concurrency.lockutils [req-b4c5f0a8-30d7-4dae-9c6b-ab2db7f66603 req-013af72a-9d3f-445c-b00d-81f8d93f24cc service nova] Acquired lock "refresh_cache-918eaae2-f821-402a-a0c2-2c140e3df4d7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1242.522902] env[63088]: DEBUG nova.network.neutron [req-b4c5f0a8-30d7-4dae-9c6b-ab2db7f66603 req-013af72a-9d3f-445c-b00d-81f8d93f24cc service nova] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] Refreshing network info cache for port c1808fff-b8ad-486e-bd90-0f23d115e398 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1242.619210] env[63088]: DEBUG oslo_concurrency.lockutils [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquiring lock "3d23868c-b0fd-470e-970c-c9935739b7b7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1242.619447] env[63088]: DEBUG oslo_concurrency.lockutils [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lock "3d23868c-b0fd-470e-970c-c9935739b7b7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1242.980360] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285631, 'name': CreateVM_Task, 'duration_secs': 0.26282} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1242.980536] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] Created VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1242.981139] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1242.981313] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1242.981627] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1242.981888] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c5b9be2b-ab0a-4487-8d6c-0d09014fd686 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.987117] env[63088]: DEBUG oslo_vmware.api [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Waiting for the task: (returnval){ [ 1242.987117] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52477908-1d72-4aa3-4b20-9554f1ce08e5" [ 1242.987117] env[63088]: _type = "Task" [ 1242.987117] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1242.995355] env[63088]: DEBUG oslo_vmware.api [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52477908-1d72-4aa3-4b20-9554f1ce08e5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1243.121804] env[63088]: DEBUG nova.compute.manager [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1243.239453] env[63088]: DEBUG nova.network.neutron [req-b4c5f0a8-30d7-4dae-9c6b-ab2db7f66603 req-013af72a-9d3f-445c-b00d-81f8d93f24cc service nova] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] Updated VIF entry in instance network info cache for port c1808fff-b8ad-486e-bd90-0f23d115e398. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1243.239865] env[63088]: DEBUG nova.network.neutron [req-b4c5f0a8-30d7-4dae-9c6b-ab2db7f66603 req-013af72a-9d3f-445c-b00d-81f8d93f24cc service nova] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] Updating instance_info_cache with network_info: [{"id": "c1808fff-b8ad-486e-bd90-0f23d115e398", "address": "fa:16:3e:15:72:9f", "network": {"id": "ab40c372-54e5-49d6-a02a-279901071b67", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1152691158-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e4d8e091fab04519a6c430290830998c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc1808fff-b8", "ovs_interfaceid": "c1808fff-b8ad-486e-bd90-0f23d115e398", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1243.497697] env[63088]: DEBUG oslo_vmware.api [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52477908-1d72-4aa3-4b20-9554f1ce08e5, 'name': SearchDatastore_Task, 'duration_secs': 0.008505} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1243.498057] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1243.498201] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] Processing image 1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1243.498432] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1243.498583] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1243.498764] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1243.499058] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-989ca460-765f-4274-85f3-7e1d032b8dd0 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.507423] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1243.507603] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1243.508275] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-484c2003-602b-445d-bc29-bfddc5ea60f0 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.512936] env[63088]: DEBUG oslo_vmware.api [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Waiting for the task: (returnval){ [ 1243.512936] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52d2710f-9f1d-de91-43fa-35164f24f92f" [ 1243.512936] env[63088]: _type = "Task" [ 1243.512936] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1243.519757] env[63088]: DEBUG oslo_vmware.api [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52d2710f-9f1d-de91-43fa-35164f24f92f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1243.642545] env[63088]: DEBUG oslo_concurrency.lockutils [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1243.642814] env[63088]: DEBUG oslo_concurrency.lockutils [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1243.644272] env[63088]: INFO nova.compute.claims [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1243.742936] env[63088]: DEBUG oslo_concurrency.lockutils [req-b4c5f0a8-30d7-4dae-9c6b-ab2db7f66603 req-013af72a-9d3f-445c-b00d-81f8d93f24cc service nova] Releasing lock "refresh_cache-918eaae2-f821-402a-a0c2-2c140e3df4d7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1244.022923] env[63088]: DEBUG oslo_vmware.api [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52d2710f-9f1d-de91-43fa-35164f24f92f, 'name': SearchDatastore_Task, 'duration_secs': 0.029953} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1244.023694] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-384be085-4f75-4fd0-9a54-104d5587f480 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.028360] env[63088]: DEBUG oslo_vmware.api [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Waiting for the task: (returnval){ [ 1244.028360] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52d210f9-7d99-ae8c-16f9-c95f51be1a15" [ 1244.028360] env[63088]: _type = "Task" [ 1244.028360] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1244.035307] env[63088]: DEBUG oslo_vmware.api [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52d210f9-7d99-ae8c-16f9-c95f51be1a15, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1244.539684] env[63088]: DEBUG oslo_vmware.api [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52d210f9-7d99-ae8c-16f9-c95f51be1a15, 'name': SearchDatastore_Task, 'duration_secs': 0.00831} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1244.540050] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1244.540163] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] 918eaae2-f821-402a-a0c2-2c140e3df4d7/918eaae2-f821-402a-a0c2-2c140e3df4d7.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1244.540430] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-aab992cf-65e5-4d74-a48d-8b280674ba43 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.546549] env[63088]: DEBUG oslo_vmware.api [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Waiting for the task: (returnval){ [ 1244.546549] env[63088]: value = "task-1285632" [ 1244.546549] env[63088]: _type = "Task" [ 1244.546549] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1244.553897] env[63088]: DEBUG oslo_vmware.api [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285632, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1244.693259] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2e3a805-381a-4cd5-b022-09e1e57bb15c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.700836] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-551bb4b0-e327-48bf-8c27-6658adcb6f76 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.731448] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c26920ee-344c-40ab-9dec-7aa051dd7d08 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.739405] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a802d507-e9fd-445b-8717-2d0ecbd90e39 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.753058] env[63088]: DEBUG nova.compute.provider_tree [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1245.056952] env[63088]: DEBUG oslo_vmware.api [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285632, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.422661} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1245.057254] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] 918eaae2-f821-402a-a0c2-2c140e3df4d7/918eaae2-f821-402a-a0c2-2c140e3df4d7.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1245.057470] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] Extending root virtual disk to 1048576 {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1245.057751] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ea9d9d70-8dc2-49f9-96b2-a45df666aaef {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.065246] env[63088]: DEBUG oslo_vmware.api [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Waiting for the task: (returnval){ [ 1245.065246] env[63088]: value = "task-1285633" [ 1245.065246] env[63088]: _type = "Task" [ 1245.065246] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1245.072995] env[63088]: DEBUG oslo_vmware.api [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285633, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1245.256403] env[63088]: DEBUG nova.scheduler.client.report [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1245.575343] env[63088]: DEBUG oslo_vmware.api [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285633, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061595} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1245.575687] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] Extended root virtual disk {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1245.576419] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0dcbc90-c14e-482f-a232-ec8b3a862308 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.597444] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] Reconfiguring VM instance instance-0000006c to attach disk [datastore1] 918eaae2-f821-402a-a0c2-2c140e3df4d7/918eaae2-f821-402a-a0c2-2c140e3df4d7.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1245.597675] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4b3d8e08-bc15-4be4-8dbf-8a8d05baa18c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.616230] env[63088]: DEBUG oslo_vmware.api [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Waiting for the task: (returnval){ [ 1245.616230] env[63088]: value = "task-1285634" [ 1245.616230] env[63088]: _type = "Task" [ 1245.616230] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1245.623355] env[63088]: DEBUG oslo_vmware.api [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285634, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1245.761552] env[63088]: DEBUG oslo_concurrency.lockutils [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.119s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1245.762163] env[63088]: DEBUG nova.compute.manager [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1246.125688] env[63088]: DEBUG oslo_vmware.api [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285634, 'name': ReconfigVM_Task, 'duration_secs': 0.248608} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1246.125973] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] Reconfigured VM instance instance-0000006c to attach disk [datastore1] 918eaae2-f821-402a-a0c2-2c140e3df4d7/918eaae2-f821-402a-a0c2-2c140e3df4d7.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1246.126590] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ac6213dc-0d92-4ac2-97d6-e7b8da17c768 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.132921] env[63088]: DEBUG oslo_vmware.api [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Waiting for the task: (returnval){ [ 1246.132921] env[63088]: value = "task-1285635" [ 1246.132921] env[63088]: _type = "Task" [ 1246.132921] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1246.139946] env[63088]: DEBUG oslo_vmware.api [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285635, 'name': Rename_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1246.267680] env[63088]: DEBUG nova.compute.utils [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1246.268958] env[63088]: DEBUG nova.compute.manager [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1246.269142] env[63088]: DEBUG nova.network.neutron [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1246.312579] env[63088]: DEBUG nova.policy [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6ffb8b92c6d649c4b39d8df1c1d0c36c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a7898ddafe0d41038a6ae4277f2c2f48', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 1246.558015] env[63088]: DEBUG nova.network.neutron [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Successfully created port: e7ade0ec-6647-4782-8915-4edcb26c64e1 {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1246.642195] env[63088]: DEBUG oslo_vmware.api [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285635, 'name': Rename_Task, 'duration_secs': 0.127377} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1246.642521] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1246.642892] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1f7fdaeb-36b8-4c2c-9869-a2e6abc204c2 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.648290] env[63088]: DEBUG oslo_vmware.api [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Waiting for the task: (returnval){ [ 1246.648290] env[63088]: value = "task-1285636" [ 1246.648290] env[63088]: _type = "Task" [ 1246.648290] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1246.655076] env[63088]: DEBUG oslo_vmware.api [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285636, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1246.772579] env[63088]: DEBUG nova.compute.manager [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1247.158246] env[63088]: DEBUG oslo_vmware.api [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285636, 'name': PowerOnVM_Task, 'duration_secs': 0.423428} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1247.158532] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1247.158742] env[63088]: INFO nova.compute.manager [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] Took 6.85 seconds to spawn the instance on the hypervisor. [ 1247.158926] env[63088]: DEBUG nova.compute.manager [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1247.159721] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e01f18c1-9eac-4c2e-bdee-1a6b93a8c713 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.677535] env[63088]: INFO nova.compute.manager [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] Took 11.53 seconds to build instance. [ 1247.785269] env[63088]: DEBUG nova.compute.manager [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1247.810888] env[63088]: DEBUG nova.virt.hardware [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1247.811241] env[63088]: DEBUG nova.virt.hardware [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1247.811420] env[63088]: DEBUG nova.virt.hardware [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1247.811609] env[63088]: DEBUG nova.virt.hardware [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1247.811764] env[63088]: DEBUG nova.virt.hardware [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1247.811915] env[63088]: DEBUG nova.virt.hardware [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1247.812158] env[63088]: DEBUG nova.virt.hardware [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1247.812335] env[63088]: DEBUG nova.virt.hardware [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1247.812505] env[63088]: DEBUG nova.virt.hardware [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1247.812674] env[63088]: DEBUG nova.virt.hardware [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1247.812847] env[63088]: DEBUG nova.virt.hardware [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1247.813803] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fa204eb-87a6-4eb0-931b-41599748c089 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.822017] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f509845d-0b1f-42fa-848c-8e265e847c00 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.923996] env[63088]: DEBUG nova.compute.manager [req-e1528bae-5cda-49e0-83cf-e9ef8ed83034 req-debfffa8-39ed-4888-aaa1-964751e71508 service nova] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Received event network-vif-plugged-e7ade0ec-6647-4782-8915-4edcb26c64e1 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1247.924229] env[63088]: DEBUG oslo_concurrency.lockutils [req-e1528bae-5cda-49e0-83cf-e9ef8ed83034 req-debfffa8-39ed-4888-aaa1-964751e71508 service nova] Acquiring lock "3d23868c-b0fd-470e-970c-c9935739b7b7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1247.924453] env[63088]: DEBUG oslo_concurrency.lockutils [req-e1528bae-5cda-49e0-83cf-e9ef8ed83034 req-debfffa8-39ed-4888-aaa1-964751e71508 service nova] Lock "3d23868c-b0fd-470e-970c-c9935739b7b7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1247.924692] env[63088]: DEBUG oslo_concurrency.lockutils [req-e1528bae-5cda-49e0-83cf-e9ef8ed83034 req-debfffa8-39ed-4888-aaa1-964751e71508 service nova] Lock "3d23868c-b0fd-470e-970c-c9935739b7b7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1247.924898] env[63088]: DEBUG nova.compute.manager [req-e1528bae-5cda-49e0-83cf-e9ef8ed83034 req-debfffa8-39ed-4888-aaa1-964751e71508 service nova] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] No waiting events found dispatching network-vif-plugged-e7ade0ec-6647-4782-8915-4edcb26c64e1 {{(pid=63088) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1247.928215] env[63088]: WARNING nova.compute.manager [req-e1528bae-5cda-49e0-83cf-e9ef8ed83034 req-debfffa8-39ed-4888-aaa1-964751e71508 service nova] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Received unexpected event network-vif-plugged-e7ade0ec-6647-4782-8915-4edcb26c64e1 for instance with vm_state building and task_state spawning. [ 1248.008853] env[63088]: DEBUG nova.network.neutron [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Successfully updated port: e7ade0ec-6647-4782-8915-4edcb26c64e1 {{(pid=63088) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1248.179319] env[63088]: DEBUG oslo_concurrency.lockutils [None req-4d3f49f1-ea46-4764-9031-c52783fa914c tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Lock "918eaae2-f821-402a-a0c2-2c140e3df4d7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.035s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1248.202952] env[63088]: DEBUG nova.compute.manager [req-8ae1dba7-95e1-4008-a155-6d70160a8dd1 req-6792999d-ca39-4acb-b16b-994ba53f01d1 service nova] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] Received event network-changed-c1808fff-b8ad-486e-bd90-0f23d115e398 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1248.202952] env[63088]: DEBUG nova.compute.manager [req-8ae1dba7-95e1-4008-a155-6d70160a8dd1 req-6792999d-ca39-4acb-b16b-994ba53f01d1 service nova] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] Refreshing instance network info cache due to event network-changed-c1808fff-b8ad-486e-bd90-0f23d115e398. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1248.203292] env[63088]: DEBUG oslo_concurrency.lockutils [req-8ae1dba7-95e1-4008-a155-6d70160a8dd1 req-6792999d-ca39-4acb-b16b-994ba53f01d1 service nova] Acquiring lock "refresh_cache-918eaae2-f821-402a-a0c2-2c140e3df4d7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1248.203292] env[63088]: DEBUG oslo_concurrency.lockutils [req-8ae1dba7-95e1-4008-a155-6d70160a8dd1 req-6792999d-ca39-4acb-b16b-994ba53f01d1 service nova] Acquired lock "refresh_cache-918eaae2-f821-402a-a0c2-2c140e3df4d7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1248.203421] env[63088]: DEBUG nova.network.neutron [req-8ae1dba7-95e1-4008-a155-6d70160a8dd1 req-6792999d-ca39-4acb-b16b-994ba53f01d1 service nova] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] Refreshing network info cache for port c1808fff-b8ad-486e-bd90-0f23d115e398 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1248.511782] env[63088]: DEBUG oslo_concurrency.lockutils [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquiring lock "refresh_cache-3d23868c-b0fd-470e-970c-c9935739b7b7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1248.511963] env[63088]: DEBUG oslo_concurrency.lockutils [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquired lock "refresh_cache-3d23868c-b0fd-470e-970c-c9935739b7b7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1248.512090] env[63088]: DEBUG nova.network.neutron [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1248.902406] env[63088]: DEBUG nova.network.neutron [req-8ae1dba7-95e1-4008-a155-6d70160a8dd1 req-6792999d-ca39-4acb-b16b-994ba53f01d1 service nova] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] Updated VIF entry in instance network info cache for port c1808fff-b8ad-486e-bd90-0f23d115e398. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1248.902802] env[63088]: DEBUG nova.network.neutron [req-8ae1dba7-95e1-4008-a155-6d70160a8dd1 req-6792999d-ca39-4acb-b16b-994ba53f01d1 service nova] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] Updating instance_info_cache with network_info: [{"id": "c1808fff-b8ad-486e-bd90-0f23d115e398", "address": "fa:16:3e:15:72:9f", "network": {"id": "ab40c372-54e5-49d6-a02a-279901071b67", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1152691158-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.133", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e4d8e091fab04519a6c430290830998c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc1808fff-b8", "ovs_interfaceid": "c1808fff-b8ad-486e-bd90-0f23d115e398", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1249.042123] env[63088]: DEBUG nova.network.neutron [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1249.158656] env[63088]: DEBUG nova.network.neutron [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Updating instance_info_cache with network_info: [{"id": "e7ade0ec-6647-4782-8915-4edcb26c64e1", "address": "fa:16:3e:43:c0:3b", "network": {"id": "711d3d9a-86db-4aa1-87b5-481f1812dfa4", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-59376186-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a7898ddafe0d41038a6ae4277f2c2f48", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e41070eb-3ac1-4ca9-a3d0-fd65893a97de", "external-id": "nsx-vlan-transportzone-596", "segmentation_id": 596, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape7ade0ec-66", "ovs_interfaceid": "e7ade0ec-6647-4782-8915-4edcb26c64e1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1249.405512] env[63088]: DEBUG oslo_concurrency.lockutils [req-8ae1dba7-95e1-4008-a155-6d70160a8dd1 req-6792999d-ca39-4acb-b16b-994ba53f01d1 service nova] Releasing lock "refresh_cache-918eaae2-f821-402a-a0c2-2c140e3df4d7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1249.662583] env[63088]: DEBUG oslo_concurrency.lockutils [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Releasing lock "refresh_cache-3d23868c-b0fd-470e-970c-c9935739b7b7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1249.662925] env[63088]: DEBUG nova.compute.manager [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Instance network_info: |[{"id": "e7ade0ec-6647-4782-8915-4edcb26c64e1", "address": "fa:16:3e:43:c0:3b", "network": {"id": "711d3d9a-86db-4aa1-87b5-481f1812dfa4", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-59376186-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a7898ddafe0d41038a6ae4277f2c2f48", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e41070eb-3ac1-4ca9-a3d0-fd65893a97de", "external-id": "nsx-vlan-transportzone-596", "segmentation_id": 596, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape7ade0ec-66", "ovs_interfaceid": "e7ade0ec-6647-4782-8915-4edcb26c64e1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1249.663396] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:43:c0:3b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e41070eb-3ac1-4ca9-a3d0-fd65893a97de', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e7ade0ec-6647-4782-8915-4edcb26c64e1', 'vif_model': 'vmxnet3'}] {{(pid=63088) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1249.671128] env[63088]: DEBUG oslo.service.loopingcall [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1249.671374] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Creating VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1249.671624] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-781761f7-b9df-46db-893d-20fce636fb36 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.691945] env[63088]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1249.691945] env[63088]: value = "task-1285637" [ 1249.691945] env[63088]: _type = "Task" [ 1249.691945] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1249.699665] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285637, 'name': CreateVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1249.950526] env[63088]: DEBUG nova.compute.manager [req-e4f3cabf-7d69-48b5-aa91-d5cb5bdccd31 req-bf4fff6e-c51b-49da-adf6-f0e34c4ef289 service nova] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Received event network-changed-e7ade0ec-6647-4782-8915-4edcb26c64e1 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1249.950777] env[63088]: DEBUG nova.compute.manager [req-e4f3cabf-7d69-48b5-aa91-d5cb5bdccd31 req-bf4fff6e-c51b-49da-adf6-f0e34c4ef289 service nova] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Refreshing instance network info cache due to event network-changed-e7ade0ec-6647-4782-8915-4edcb26c64e1. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1249.951053] env[63088]: DEBUG oslo_concurrency.lockutils [req-e4f3cabf-7d69-48b5-aa91-d5cb5bdccd31 req-bf4fff6e-c51b-49da-adf6-f0e34c4ef289 service nova] Acquiring lock "refresh_cache-3d23868c-b0fd-470e-970c-c9935739b7b7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1249.951255] env[63088]: DEBUG oslo_concurrency.lockutils [req-e4f3cabf-7d69-48b5-aa91-d5cb5bdccd31 req-bf4fff6e-c51b-49da-adf6-f0e34c4ef289 service nova] Acquired lock "refresh_cache-3d23868c-b0fd-470e-970c-c9935739b7b7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1249.951473] env[63088]: DEBUG nova.network.neutron [req-e4f3cabf-7d69-48b5-aa91-d5cb5bdccd31 req-bf4fff6e-c51b-49da-adf6-f0e34c4ef289 service nova] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Refreshing network info cache for port e7ade0ec-6647-4782-8915-4edcb26c64e1 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1250.201991] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285637, 'name': CreateVM_Task, 'duration_secs': 0.28831} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1250.202193] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Created VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1250.202773] env[63088]: DEBUG oslo_concurrency.lockutils [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1250.202958] env[63088]: DEBUG oslo_concurrency.lockutils [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1250.203291] env[63088]: DEBUG oslo_concurrency.lockutils [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1250.203544] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d77f6b3f-e147-47e1-8c70-07b17df26446 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.207654] env[63088]: DEBUG oslo_vmware.api [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1250.207654] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]5269c6de-9310-e6e0-e051-c07bf300194e" [ 1250.207654] env[63088]: _type = "Task" [ 1250.207654] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1250.214932] env[63088]: DEBUG oslo_vmware.api [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]5269c6de-9310-e6e0-e051-c07bf300194e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1250.625178] env[63088]: DEBUG nova.network.neutron [req-e4f3cabf-7d69-48b5-aa91-d5cb5bdccd31 req-bf4fff6e-c51b-49da-adf6-f0e34c4ef289 service nova] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Updated VIF entry in instance network info cache for port e7ade0ec-6647-4782-8915-4edcb26c64e1. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1250.625643] env[63088]: DEBUG nova.network.neutron [req-e4f3cabf-7d69-48b5-aa91-d5cb5bdccd31 req-bf4fff6e-c51b-49da-adf6-f0e34c4ef289 service nova] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Updating instance_info_cache with network_info: [{"id": "e7ade0ec-6647-4782-8915-4edcb26c64e1", "address": "fa:16:3e:43:c0:3b", "network": {"id": "711d3d9a-86db-4aa1-87b5-481f1812dfa4", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-59376186-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a7898ddafe0d41038a6ae4277f2c2f48", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e41070eb-3ac1-4ca9-a3d0-fd65893a97de", "external-id": "nsx-vlan-transportzone-596", "segmentation_id": 596, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape7ade0ec-66", "ovs_interfaceid": "e7ade0ec-6647-4782-8915-4edcb26c64e1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1250.718134] env[63088]: DEBUG oslo_vmware.api [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]5269c6de-9310-e6e0-e051-c07bf300194e, 'name': SearchDatastore_Task, 'duration_secs': 0.009577} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1250.718520] env[63088]: DEBUG oslo_concurrency.lockutils [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1250.718826] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Processing image 1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1250.719093] env[63088]: DEBUG oslo_concurrency.lockutils [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1250.719250] env[63088]: DEBUG oslo_concurrency.lockutils [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1250.719433] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1250.719785] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3783ecff-0f18-40db-9fab-0154ca80c3cc {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.728916] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1250.729123] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1250.729843] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0764042f-b1ff-4ce8-b9bc-3cb769b0c6e9 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.734985] env[63088]: DEBUG oslo_vmware.api [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1250.734985] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]525dc1d6-ea79-004d-c54e-de86f721d30b" [ 1250.734985] env[63088]: _type = "Task" [ 1250.734985] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1250.742842] env[63088]: DEBUG oslo_vmware.api [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]525dc1d6-ea79-004d-c54e-de86f721d30b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1251.128873] env[63088]: DEBUG oslo_concurrency.lockutils [req-e4f3cabf-7d69-48b5-aa91-d5cb5bdccd31 req-bf4fff6e-c51b-49da-adf6-f0e34c4ef289 service nova] Releasing lock "refresh_cache-3d23868c-b0fd-470e-970c-c9935739b7b7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1251.246562] env[63088]: DEBUG oslo_vmware.api [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]525dc1d6-ea79-004d-c54e-de86f721d30b, 'name': SearchDatastore_Task, 'duration_secs': 0.008001} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1251.247392] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9821c5e8-bd60-4ee9-95c4-d43e8a36191f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.252563] env[63088]: DEBUG oslo_vmware.api [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1251.252563] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]529615cc-7619-7ffb-a17f-fa1f08b1b2f1" [ 1251.252563] env[63088]: _type = "Task" [ 1251.252563] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1251.260262] env[63088]: DEBUG oslo_vmware.api [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]529615cc-7619-7ffb-a17f-fa1f08b1b2f1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1251.763048] env[63088]: DEBUG oslo_vmware.api [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]529615cc-7619-7ffb-a17f-fa1f08b1b2f1, 'name': SearchDatastore_Task, 'duration_secs': 0.00976} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1251.763284] env[63088]: DEBUG oslo_concurrency.lockutils [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1251.763548] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] 3d23868c-b0fd-470e-970c-c9935739b7b7/3d23868c-b0fd-470e-970c-c9935739b7b7.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1251.763806] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d6766159-096e-4623-9994-a88d16e77787 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.770177] env[63088]: DEBUG oslo_vmware.api [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1251.770177] env[63088]: value = "task-1285638" [ 1251.770177] env[63088]: _type = "Task" [ 1251.770177] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1251.777387] env[63088]: DEBUG oslo_vmware.api [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285638, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1252.279724] env[63088]: DEBUG oslo_vmware.api [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285638, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.463618} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1252.280127] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] 3d23868c-b0fd-470e-970c-c9935739b7b7/3d23868c-b0fd-470e-970c-c9935739b7b7.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1252.280206] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Extending root virtual disk to 1048576 {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1252.280444] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b521cf4a-6cd4-408a-8436-2aff47145cee {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.286733] env[63088]: DEBUG oslo_vmware.api [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1252.286733] env[63088]: value = "task-1285639" [ 1252.286733] env[63088]: _type = "Task" [ 1252.286733] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1252.294206] env[63088]: DEBUG oslo_vmware.api [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285639, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1252.796231] env[63088]: DEBUG oslo_vmware.api [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285639, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.059912} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1252.796499] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Extended root virtual disk {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1252.797261] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca6eab45-97da-4570-b848-4c0ceaf450f4 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.818374] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Reconfiguring VM instance instance-0000006d to attach disk [datastore1] 3d23868c-b0fd-470e-970c-c9935739b7b7/3d23868c-b0fd-470e-970c-c9935739b7b7.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1252.818619] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f6dd0498-4161-43e0-a1c7-b96b5c5a75c6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.837435] env[63088]: DEBUG oslo_vmware.api [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1252.837435] env[63088]: value = "task-1285640" [ 1252.837435] env[63088]: _type = "Task" [ 1252.837435] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1252.844500] env[63088]: DEBUG oslo_vmware.api [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285640, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1253.346944] env[63088]: DEBUG oslo_vmware.api [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285640, 'name': ReconfigVM_Task, 'duration_secs': 0.261954} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1253.348058] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Reconfigured VM instance instance-0000006d to attach disk [datastore1] 3d23868c-b0fd-470e-970c-c9935739b7b7/3d23868c-b0fd-470e-970c-c9935739b7b7.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1253.348386] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-42311ed0-6fb4-45b8-8a79-602e441a254e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.355379] env[63088]: DEBUG oslo_vmware.api [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1253.355379] env[63088]: value = "task-1285641" [ 1253.355379] env[63088]: _type = "Task" [ 1253.355379] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1253.364174] env[63088]: DEBUG oslo_vmware.api [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285641, 'name': Rename_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1253.864881] env[63088]: DEBUG oslo_vmware.api [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285641, 'name': Rename_Task, 'duration_secs': 0.175229} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1253.865180] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1253.865433] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1f582a18-47e8-4869-83c7-95bbe411cb3d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.871504] env[63088]: DEBUG oslo_vmware.api [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1253.871504] env[63088]: value = "task-1285642" [ 1253.871504] env[63088]: _type = "Task" [ 1253.871504] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1253.878571] env[63088]: DEBUG oslo_vmware.api [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285642, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1254.380857] env[63088]: DEBUG oslo_vmware.api [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285642, 'name': PowerOnVM_Task, 'duration_secs': 0.48876} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1254.381238] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1254.381361] env[63088]: INFO nova.compute.manager [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Took 6.60 seconds to spawn the instance on the hypervisor. [ 1254.381526] env[63088]: DEBUG nova.compute.manager [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1254.382316] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d242d2ad-e2c7-4c1a-b223-911cc042a65b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.898835] env[63088]: INFO nova.compute.manager [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Took 11.27 seconds to build instance. [ 1255.400171] env[63088]: DEBUG oslo_concurrency.lockutils [None req-618dcc50-9422-4061-a590-b9e171e5af62 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lock "3d23868c-b0fd-470e-970c-c9935739b7b7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.780s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1255.679619] env[63088]: DEBUG nova.compute.manager [req-ddc88831-c2b0-41df-b664-21c176c36c2d req-e481e546-ea43-4ee0-ac8b-35ff52388527 service nova] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Received event network-changed-e7ade0ec-6647-4782-8915-4edcb26c64e1 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1255.679907] env[63088]: DEBUG nova.compute.manager [req-ddc88831-c2b0-41df-b664-21c176c36c2d req-e481e546-ea43-4ee0-ac8b-35ff52388527 service nova] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Refreshing instance network info cache due to event network-changed-e7ade0ec-6647-4782-8915-4edcb26c64e1. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1255.680244] env[63088]: DEBUG oslo_concurrency.lockutils [req-ddc88831-c2b0-41df-b664-21c176c36c2d req-e481e546-ea43-4ee0-ac8b-35ff52388527 service nova] Acquiring lock "refresh_cache-3d23868c-b0fd-470e-970c-c9935739b7b7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1255.680472] env[63088]: DEBUG oslo_concurrency.lockutils [req-ddc88831-c2b0-41df-b664-21c176c36c2d req-e481e546-ea43-4ee0-ac8b-35ff52388527 service nova] Acquired lock "refresh_cache-3d23868c-b0fd-470e-970c-c9935739b7b7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1255.680728] env[63088]: DEBUG nova.network.neutron [req-ddc88831-c2b0-41df-b664-21c176c36c2d req-e481e546-ea43-4ee0-ac8b-35ff52388527 service nova] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Refreshing network info cache for port e7ade0ec-6647-4782-8915-4edcb26c64e1 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1256.382732] env[63088]: DEBUG nova.network.neutron [req-ddc88831-c2b0-41df-b664-21c176c36c2d req-e481e546-ea43-4ee0-ac8b-35ff52388527 service nova] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Updated VIF entry in instance network info cache for port e7ade0ec-6647-4782-8915-4edcb26c64e1. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1256.383133] env[63088]: DEBUG nova.network.neutron [req-ddc88831-c2b0-41df-b664-21c176c36c2d req-e481e546-ea43-4ee0-ac8b-35ff52388527 service nova] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Updating instance_info_cache with network_info: [{"id": "e7ade0ec-6647-4782-8915-4edcb26c64e1", "address": "fa:16:3e:43:c0:3b", "network": {"id": "711d3d9a-86db-4aa1-87b5-481f1812dfa4", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-59376186-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.225", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a7898ddafe0d41038a6ae4277f2c2f48", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e41070eb-3ac1-4ca9-a3d0-fd65893a97de", "external-id": "nsx-vlan-transportzone-596", "segmentation_id": 596, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape7ade0ec-66", "ovs_interfaceid": "e7ade0ec-6647-4782-8915-4edcb26c64e1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1256.885950] env[63088]: DEBUG oslo_concurrency.lockutils [req-ddc88831-c2b0-41df-b664-21c176c36c2d req-e481e546-ea43-4ee0-ac8b-35ff52388527 service nova] Releasing lock "refresh_cache-3d23868c-b0fd-470e-970c-c9935739b7b7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1274.604315] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1278.605053] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1278.605399] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1280.604243] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1280.604628] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63088) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1282.604489] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1282.604769] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Starting heal instance info cache {{(pid=63088) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1282.604769] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Rebuilding the list of instances to heal {{(pid=63088) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 1283.137391] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Acquiring lock "refresh_cache-918eaae2-f821-402a-a0c2-2c140e3df4d7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1283.137537] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Acquired lock "refresh_cache-918eaae2-f821-402a-a0c2-2c140e3df4d7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1283.137685] env[63088]: DEBUG nova.network.neutron [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] Forcefully refreshing network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1283.137847] env[63088]: DEBUG nova.objects.instance [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Lazy-loading 'info_cache' on Instance uuid 918eaae2-f821-402a-a0c2-2c140e3df4d7 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1284.847207] env[63088]: DEBUG nova.network.neutron [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] Updating instance_info_cache with network_info: [{"id": "c1808fff-b8ad-486e-bd90-0f23d115e398", "address": "fa:16:3e:15:72:9f", "network": {"id": "ab40c372-54e5-49d6-a02a-279901071b67", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1152691158-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.133", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e4d8e091fab04519a6c430290830998c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc1808fff-b8", "ovs_interfaceid": "c1808fff-b8ad-486e-bd90-0f23d115e398", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1285.349690] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Releasing lock "refresh_cache-918eaae2-f821-402a-a0c2-2c140e3df4d7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1285.349893] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] Updated the network info_cache for instance {{(pid=63088) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 1285.350248] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1285.350430] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1285.350585] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager.update_available_resource {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1285.853072] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1285.853445] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1285.853445] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1285.853594] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63088) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1285.854600] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e11f0180-4629-4d29-8ebe-04dcf25c8968 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1285.863353] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7d7971b-1481-4a41-86c2-9d224ae93964 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1286.536745] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de632103-0b08-4207-81ee-9afff4d5ea96 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1286.541100] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5217893c-8f37-4c5d-b46f-6aaec5e84ec6 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Acquiring lock "918eaae2-f821-402a-a0c2-2c140e3df4d7" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1286.541331] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5217893c-8f37-4c5d-b46f-6aaec5e84ec6 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Lock "918eaae2-f821-402a-a0c2-2c140e3df4d7" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1286.545255] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16330ddd-25e6-45af-bbc5-984e483827da {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1286.573147] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181320MB free_disk=140GB free_vcpus=48 pci_devices=None {{(pid=63088) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1286.573292] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1286.573463] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1287.046416] env[63088]: DEBUG nova.compute.utils [None req-5217893c-8f37-4c5d-b46f-6aaec5e84ec6 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1287.549584] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5217893c-8f37-4c5d-b46f-6aaec5e84ec6 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Lock "918eaae2-f821-402a-a0c2-2c140e3df4d7" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.008s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1287.598164] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance 918eaae2-f821-402a-a0c2-2c140e3df4d7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1287.598339] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance 3d23868c-b0fd-470e-970c-c9935739b7b7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1287.598514] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Total usable vcpus: 48, total allocated vcpus: 2 {{(pid=63088) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1287.598653] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=896MB phys_disk=200GB used_disk=2GB total_vcpus=48 used_vcpus=2 pci_stats=[] {{(pid=63088) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1287.635311] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-265ffc29-d79c-4cc2-89e0-1ed45bd2df20 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1287.642804] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7be56856-bf00-49fe-894a-d8aaf4cc5526 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1287.671871] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47765c96-5b6c-4f61-a0d7-c6acc457732e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1287.678299] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbed0937-03f6-48dd-bc69-a17d0d970f8f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1287.690460] env[63088]: DEBUG nova.compute.provider_tree [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1288.193690] env[63088]: DEBUG nova.scheduler.client.report [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1288.432905] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5217893c-8f37-4c5d-b46f-6aaec5e84ec6 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Acquiring lock "918eaae2-f821-402a-a0c2-2c140e3df4d7" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1288.433250] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5217893c-8f37-4c5d-b46f-6aaec5e84ec6 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Lock "918eaae2-f821-402a-a0c2-2c140e3df4d7" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1288.433435] env[63088]: INFO nova.compute.manager [None req-5217893c-8f37-4c5d-b46f-6aaec5e84ec6 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] Attaching volume 6bdbf370-3ba7-475f-b2c7-d60435f73bd7 to /dev/sdb [ 1288.463790] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2ee0815-5921-428b-ab95-de9a303b6948 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1288.470963] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33b7f33e-e63a-4aa6-87d6-5d44eccc9a86 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1288.483403] env[63088]: DEBUG nova.virt.block_device [None req-5217893c-8f37-4c5d-b46f-6aaec5e84ec6 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] Updating existing volume attachment record: 6bd3b703-9b59-48e4-9850-97211b7584c5 {{(pid=63088) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1288.698439] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63088) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1288.698650] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.125s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1289.694997] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1293.026761] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-5217893c-8f37-4c5d-b46f-6aaec5e84ec6 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] Volume attach. Driver type: vmdk {{(pid=63088) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1293.027051] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-5217893c-8f37-4c5d-b46f-6aaec5e84ec6 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-275980', 'volume_id': '6bdbf370-3ba7-475f-b2c7-d60435f73bd7', 'name': 'volume-6bdbf370-3ba7-475f-b2c7-d60435f73bd7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '918eaae2-f821-402a-a0c2-2c140e3df4d7', 'attached_at': '', 'detached_at': '', 'volume_id': '6bdbf370-3ba7-475f-b2c7-d60435f73bd7', 'serial': '6bdbf370-3ba7-475f-b2c7-d60435f73bd7'} {{(pid=63088) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1293.027943] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-372f866d-6714-4762-98d1-7f0587f8a244 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1293.045365] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81160dfc-4836-489a-b344-0711873c2808 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1293.068773] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-5217893c-8f37-4c5d-b46f-6aaec5e84ec6 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] Reconfiguring VM instance instance-0000006c to attach disk [datastore1] volume-6bdbf370-3ba7-475f-b2c7-d60435f73bd7/volume-6bdbf370-3ba7-475f-b2c7-d60435f73bd7.vmdk or device None with type thin {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1293.068957] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d2a6886e-374c-47dc-bf03-59c621063eff {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1293.086233] env[63088]: DEBUG oslo_vmware.api [None req-5217893c-8f37-4c5d-b46f-6aaec5e84ec6 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Waiting for the task: (returnval){ [ 1293.086233] env[63088]: value = "task-1285647" [ 1293.086233] env[63088]: _type = "Task" [ 1293.086233] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1293.093495] env[63088]: DEBUG oslo_vmware.api [None req-5217893c-8f37-4c5d-b46f-6aaec5e84ec6 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285647, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1293.404496] env[63088]: DEBUG oslo_concurrency.lockutils [None req-1160ba93-80cb-4c91-924c-c9b9da607442 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquiring lock "3d23868c-b0fd-470e-970c-c9935739b7b7" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1293.404783] env[63088]: DEBUG oslo_concurrency.lockutils [None req-1160ba93-80cb-4c91-924c-c9b9da607442 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lock "3d23868c-b0fd-470e-970c-c9935739b7b7" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1293.404984] env[63088]: INFO nova.compute.manager [None req-1160ba93-80cb-4c91-924c-c9b9da607442 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Shelving [ 1293.595642] env[63088]: DEBUG oslo_vmware.api [None req-5217893c-8f37-4c5d-b46f-6aaec5e84ec6 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285647, 'name': ReconfigVM_Task, 'duration_secs': 0.316738} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1293.595941] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-5217893c-8f37-4c5d-b46f-6aaec5e84ec6 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] Reconfigured VM instance instance-0000006c to attach disk [datastore1] volume-6bdbf370-3ba7-475f-b2c7-d60435f73bd7/volume-6bdbf370-3ba7-475f-b2c7-d60435f73bd7.vmdk or device None with type thin {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1293.600580] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-719ae215-6403-46f3-87d3-a487d502c900 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1293.615153] env[63088]: DEBUG oslo_vmware.api [None req-5217893c-8f37-4c5d-b46f-6aaec5e84ec6 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Waiting for the task: (returnval){ [ 1293.615153] env[63088]: value = "task-1285648" [ 1293.615153] env[63088]: _type = "Task" [ 1293.615153] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1293.622311] env[63088]: DEBUG oslo_vmware.api [None req-5217893c-8f37-4c5d-b46f-6aaec5e84ec6 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285648, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1293.912165] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-1160ba93-80cb-4c91-924c-c9b9da607442 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1293.912473] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-425cce39-b0fe-4ce0-a3e7-8e45b735e6bb {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1293.919591] env[63088]: DEBUG oslo_vmware.api [None req-1160ba93-80cb-4c91-924c-c9b9da607442 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1293.919591] env[63088]: value = "task-1285649" [ 1293.919591] env[63088]: _type = "Task" [ 1293.919591] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1293.927667] env[63088]: DEBUG oslo_vmware.api [None req-1160ba93-80cb-4c91-924c-c9b9da607442 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285649, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1294.124570] env[63088]: DEBUG oslo_vmware.api [None req-5217893c-8f37-4c5d-b46f-6aaec5e84ec6 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285648, 'name': ReconfigVM_Task, 'duration_secs': 0.119911} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1294.124908] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-5217893c-8f37-4c5d-b46f-6aaec5e84ec6 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-275980', 'volume_id': '6bdbf370-3ba7-475f-b2c7-d60435f73bd7', 'name': 'volume-6bdbf370-3ba7-475f-b2c7-d60435f73bd7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '918eaae2-f821-402a-a0c2-2c140e3df4d7', 'attached_at': '', 'detached_at': '', 'volume_id': '6bdbf370-3ba7-475f-b2c7-d60435f73bd7', 'serial': '6bdbf370-3ba7-475f-b2c7-d60435f73bd7'} {{(pid=63088) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1294.429123] env[63088]: DEBUG oslo_vmware.api [None req-1160ba93-80cb-4c91-924c-c9b9da607442 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285649, 'name': PowerOffVM_Task, 'duration_secs': 0.166088} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1294.429352] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-1160ba93-80cb-4c91-924c-c9b9da607442 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1294.430111] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b012ab32-60d6-48d8-a210-95dc17a53785 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1294.447030] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00c38230-e3be-4187-9c37-feb2bd4fa496 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1294.957762] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-1160ba93-80cb-4c91-924c-c9b9da607442 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Creating Snapshot of the VM instance {{(pid=63088) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1294.958251] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-214e6efa-d81f-4f42-9c45-afa18c007f26 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1294.966745] env[63088]: DEBUG oslo_vmware.api [None req-1160ba93-80cb-4c91-924c-c9b9da607442 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1294.966745] env[63088]: value = "task-1285650" [ 1294.966745] env[63088]: _type = "Task" [ 1294.966745] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1294.980131] env[63088]: DEBUG oslo_vmware.api [None req-1160ba93-80cb-4c91-924c-c9b9da607442 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285650, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1295.158727] env[63088]: DEBUG nova.objects.instance [None req-5217893c-8f37-4c5d-b46f-6aaec5e84ec6 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Lazy-loading 'flavor' on Instance uuid 918eaae2-f821-402a-a0c2-2c140e3df4d7 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1295.477344] env[63088]: DEBUG oslo_vmware.api [None req-1160ba93-80cb-4c91-924c-c9b9da607442 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285650, 'name': CreateSnapshot_Task, 'duration_secs': 0.412063} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1295.477344] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-1160ba93-80cb-4c91-924c-c9b9da607442 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Created Snapshot of the VM instance {{(pid=63088) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1295.477344] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f988e201-7d5e-4ed5-a3d4-8c5c4ff826f5 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1295.664119] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5217893c-8f37-4c5d-b46f-6aaec5e84ec6 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Lock "918eaae2-f821-402a-a0c2-2c140e3df4d7" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.231s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1295.993755] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-1160ba93-80cb-4c91-924c-c9b9da607442 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Creating linked-clone VM from snapshot {{(pid=63088) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1295.994179] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-90485370-2132-4bd8-aed7-c5fe6b93b2ff {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1296.003132] env[63088]: DEBUG oslo_vmware.api [None req-1160ba93-80cb-4c91-924c-c9b9da607442 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1296.003132] env[63088]: value = "task-1285651" [ 1296.003132] env[63088]: _type = "Task" [ 1296.003132] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1296.011309] env[63088]: DEBUG oslo_vmware.api [None req-1160ba93-80cb-4c91-924c-c9b9da607442 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285651, 'name': CloneVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1296.513546] env[63088]: DEBUG oslo_vmware.api [None req-1160ba93-80cb-4c91-924c-c9b9da607442 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285651, 'name': CloneVM_Task} progress is 94%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1296.738253] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Acquiring lock "bd55e4b9-269f-4ca4-8139-86dface352b0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1296.738494] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Lock "bd55e4b9-269f-4ca4-8139-86dface352b0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1297.016898] env[63088]: DEBUG oslo_vmware.api [None req-1160ba93-80cb-4c91-924c-c9b9da607442 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285651, 'name': CloneVM_Task, 'duration_secs': 0.901491} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1297.016898] env[63088]: INFO nova.virt.vmwareapi.vmops [None req-1160ba93-80cb-4c91-924c-c9b9da607442 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Created linked-clone VM from snapshot [ 1297.017679] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f0a6bb2-47d7-4a18-bb2d-070a2a506716 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1297.024982] env[63088]: DEBUG nova.virt.vmwareapi.images [None req-1160ba93-80cb-4c91-924c-c9b9da607442 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Uploading image 5bb4b00f-be25-4207-8793-a8341056e087 {{(pid=63088) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1297.048092] env[63088]: DEBUG oslo_vmware.rw_handles [None req-1160ba93-80cb-4c91-924c-c9b9da607442 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1297.048092] env[63088]: value = "vm-275982" [ 1297.048092] env[63088]: _type = "VirtualMachine" [ 1297.048092] env[63088]: }. {{(pid=63088) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1297.048349] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-4afc8456-cbf5-4ac8-9aaf-966c3bd50277 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1297.054164] env[63088]: DEBUG oslo_vmware.rw_handles [None req-1160ba93-80cb-4c91-924c-c9b9da607442 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lease: (returnval){ [ 1297.054164] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]528f694d-09db-3608-4a33-86544c185908" [ 1297.054164] env[63088]: _type = "HttpNfcLease" [ 1297.054164] env[63088]: } obtained for exporting VM: (result){ [ 1297.054164] env[63088]: value = "vm-275982" [ 1297.054164] env[63088]: _type = "VirtualMachine" [ 1297.054164] env[63088]: }. {{(pid=63088) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1297.054407] env[63088]: DEBUG oslo_vmware.api [None req-1160ba93-80cb-4c91-924c-c9b9da607442 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the lease: (returnval){ [ 1297.054407] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]528f694d-09db-3608-4a33-86544c185908" [ 1297.054407] env[63088]: _type = "HttpNfcLease" [ 1297.054407] env[63088]: } to be ready. {{(pid=63088) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1297.060216] env[63088]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1297.060216] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]528f694d-09db-3608-4a33-86544c185908" [ 1297.060216] env[63088]: _type = "HttpNfcLease" [ 1297.060216] env[63088]: } is initializing. {{(pid=63088) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1297.241060] env[63088]: DEBUG nova.compute.manager [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: bd55e4b9-269f-4ca4-8139-86dface352b0] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1297.561650] env[63088]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1297.561650] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]528f694d-09db-3608-4a33-86544c185908" [ 1297.561650] env[63088]: _type = "HttpNfcLease" [ 1297.561650] env[63088]: } is ready. {{(pid=63088) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1297.562056] env[63088]: DEBUG oslo_vmware.rw_handles [None req-1160ba93-80cb-4c91-924c-c9b9da607442 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1297.562056] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]528f694d-09db-3608-4a33-86544c185908" [ 1297.562056] env[63088]: _type = "HttpNfcLease" [ 1297.562056] env[63088]: }. {{(pid=63088) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1297.562630] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53c7e1cd-8456-4b83-bfa1-f7fae0e6bdfd {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1297.569420] env[63088]: DEBUG oslo_vmware.rw_handles [None req-1160ba93-80cb-4c91-924c-c9b9da607442 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ab441d-04af-3a00-6c14-bcb020137cd5/disk-0.vmdk from lease info. {{(pid=63088) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1297.569591] env[63088]: DEBUG oslo_vmware.rw_handles [None req-1160ba93-80cb-4c91-924c-c9b9da607442 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ab441d-04af-3a00-6c14-bcb020137cd5/disk-0.vmdk for reading. {{(pid=63088) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1297.654465] env[63088]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-335992bc-355b-4bbf-b879-19ee7fd42917 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1297.763901] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1297.764166] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1297.765775] env[63088]: INFO nova.compute.claims [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: bd55e4b9-269f-4ca4-8139-86dface352b0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1298.822223] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1db8972b-3787-4af6-8695-43ab6bfd6845 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1298.830649] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1af8242-9f57-4ba2-8375-c685064c82f3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1298.864020] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7274e854-5931-4a99-a3dd-86cfc1fb7527 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1298.871844] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8785357f-23d4-49dd-a44d-e4066509d28e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1298.888431] env[63088]: DEBUG nova.compute.provider_tree [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1299.393031] env[63088]: DEBUG nova.scheduler.client.report [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1299.898880] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.134s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1299.899393] env[63088]: DEBUG nova.compute.manager [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: bd55e4b9-269f-4ca4-8139-86dface352b0] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1300.404494] env[63088]: DEBUG nova.compute.utils [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1300.406437] env[63088]: DEBUG nova.compute.manager [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: bd55e4b9-269f-4ca4-8139-86dface352b0] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1300.406696] env[63088]: DEBUG nova.network.neutron [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: bd55e4b9-269f-4ca4-8139-86dface352b0] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1300.445654] env[63088]: DEBUG nova.policy [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '11a98d4e5ef3402b99bea85a0bb479e7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e4d8e091fab04519a6c430290830998c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 1300.730517] env[63088]: DEBUG nova.network.neutron [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: bd55e4b9-269f-4ca4-8139-86dface352b0] Successfully created port: b40212cc-273b-4dbb-9363-91876eec40ad {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1300.910475] env[63088]: DEBUG nova.compute.manager [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: bd55e4b9-269f-4ca4-8139-86dface352b0] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1301.920893] env[63088]: DEBUG nova.compute.manager [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: bd55e4b9-269f-4ca4-8139-86dface352b0] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1301.946375] env[63088]: DEBUG nova.virt.hardware [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1301.946712] env[63088]: DEBUG nova.virt.hardware [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1301.946895] env[63088]: DEBUG nova.virt.hardware [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1301.947103] env[63088]: DEBUG nova.virt.hardware [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1301.947262] env[63088]: DEBUG nova.virt.hardware [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1301.947429] env[63088]: DEBUG nova.virt.hardware [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1301.947635] env[63088]: DEBUG nova.virt.hardware [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1301.947801] env[63088]: DEBUG nova.virt.hardware [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1301.947975] env[63088]: DEBUG nova.virt.hardware [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1301.948161] env[63088]: DEBUG nova.virt.hardware [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1301.948342] env[63088]: DEBUG nova.virt.hardware [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1301.949232] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a6ca2e0-bf34-49e6-96f6-3f0345f105ab {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1301.957671] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7d9b4e4-bfb7-4c6f-9c9a-f7748b574efe {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1302.125403] env[63088]: DEBUG nova.compute.manager [req-faac0931-a4c7-4e62-a96c-d4a8e4180e54 req-5906972c-1e42-4d7b-add8-9b6bd25b2ac2 service nova] [instance: bd55e4b9-269f-4ca4-8139-86dface352b0] Received event network-vif-plugged-b40212cc-273b-4dbb-9363-91876eec40ad {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1302.125646] env[63088]: DEBUG oslo_concurrency.lockutils [req-faac0931-a4c7-4e62-a96c-d4a8e4180e54 req-5906972c-1e42-4d7b-add8-9b6bd25b2ac2 service nova] Acquiring lock "bd55e4b9-269f-4ca4-8139-86dface352b0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1302.125830] env[63088]: DEBUG oslo_concurrency.lockutils [req-faac0931-a4c7-4e62-a96c-d4a8e4180e54 req-5906972c-1e42-4d7b-add8-9b6bd25b2ac2 service nova] Lock "bd55e4b9-269f-4ca4-8139-86dface352b0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1302.126009] env[63088]: DEBUG oslo_concurrency.lockutils [req-faac0931-a4c7-4e62-a96c-d4a8e4180e54 req-5906972c-1e42-4d7b-add8-9b6bd25b2ac2 service nova] Lock "bd55e4b9-269f-4ca4-8139-86dface352b0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1302.127576] env[63088]: DEBUG nova.compute.manager [req-faac0931-a4c7-4e62-a96c-d4a8e4180e54 req-5906972c-1e42-4d7b-add8-9b6bd25b2ac2 service nova] [instance: bd55e4b9-269f-4ca4-8139-86dface352b0] No waiting events found dispatching network-vif-plugged-b40212cc-273b-4dbb-9363-91876eec40ad {{(pid=63088) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1302.127775] env[63088]: WARNING nova.compute.manager [req-faac0931-a4c7-4e62-a96c-d4a8e4180e54 req-5906972c-1e42-4d7b-add8-9b6bd25b2ac2 service nova] [instance: bd55e4b9-269f-4ca4-8139-86dface352b0] Received unexpected event network-vif-plugged-b40212cc-273b-4dbb-9363-91876eec40ad for instance with vm_state building and task_state spawning. [ 1302.214299] env[63088]: DEBUG nova.network.neutron [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: bd55e4b9-269f-4ca4-8139-86dface352b0] Successfully updated port: b40212cc-273b-4dbb-9363-91876eec40ad {{(pid=63088) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1302.717149] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Acquiring lock "refresh_cache-bd55e4b9-269f-4ca4-8139-86dface352b0" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1302.717372] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Acquired lock "refresh_cache-bd55e4b9-269f-4ca4-8139-86dface352b0" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1302.717601] env[63088]: DEBUG nova.network.neutron [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: bd55e4b9-269f-4ca4-8139-86dface352b0] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1303.252142] env[63088]: DEBUG nova.network.neutron [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: bd55e4b9-269f-4ca4-8139-86dface352b0] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1303.385262] env[63088]: DEBUG nova.network.neutron [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: bd55e4b9-269f-4ca4-8139-86dface352b0] Updating instance_info_cache with network_info: [{"id": "b40212cc-273b-4dbb-9363-91876eec40ad", "address": "fa:16:3e:36:62:93", "network": {"id": "ab40c372-54e5-49d6-a02a-279901071b67", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1152691158-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e4d8e091fab04519a6c430290830998c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb40212cc-27", "ovs_interfaceid": "b40212cc-273b-4dbb-9363-91876eec40ad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1303.888605] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Releasing lock "refresh_cache-bd55e4b9-269f-4ca4-8139-86dface352b0" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1303.888944] env[63088]: DEBUG nova.compute.manager [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: bd55e4b9-269f-4ca4-8139-86dface352b0] Instance network_info: |[{"id": "b40212cc-273b-4dbb-9363-91876eec40ad", "address": "fa:16:3e:36:62:93", "network": {"id": "ab40c372-54e5-49d6-a02a-279901071b67", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1152691158-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e4d8e091fab04519a6c430290830998c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb40212cc-27", "ovs_interfaceid": "b40212cc-273b-4dbb-9363-91876eec40ad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1303.889431] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: bd55e4b9-269f-4ca4-8139-86dface352b0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:36:62:93', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '418ddd3d-5f64-407e-8e0c-c8b81639bee9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b40212cc-273b-4dbb-9363-91876eec40ad', 'vif_model': 'vmxnet3'}] {{(pid=63088) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1303.897482] env[63088]: DEBUG oslo.service.loopingcall [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1303.897712] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bd55e4b9-269f-4ca4-8139-86dface352b0] Creating VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1303.899135] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ca342f77-3d61-431b-a4bd-ef2588190807 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1303.919646] env[63088]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1303.919646] env[63088]: value = "task-1285653" [ 1303.919646] env[63088]: _type = "Task" [ 1303.919646] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1303.927305] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285653, 'name': CreateVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1304.157196] env[63088]: DEBUG nova.compute.manager [req-faf61559-e210-4983-94b5-4e071ac273d7 req-8f3ff297-7c9e-4a2a-a232-bb017e47926e service nova] [instance: bd55e4b9-269f-4ca4-8139-86dface352b0] Received event network-changed-b40212cc-273b-4dbb-9363-91876eec40ad {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1304.157531] env[63088]: DEBUG nova.compute.manager [req-faf61559-e210-4983-94b5-4e071ac273d7 req-8f3ff297-7c9e-4a2a-a232-bb017e47926e service nova] [instance: bd55e4b9-269f-4ca4-8139-86dface352b0] Refreshing instance network info cache due to event network-changed-b40212cc-273b-4dbb-9363-91876eec40ad. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1304.157884] env[63088]: DEBUG oslo_concurrency.lockutils [req-faf61559-e210-4983-94b5-4e071ac273d7 req-8f3ff297-7c9e-4a2a-a232-bb017e47926e service nova] Acquiring lock "refresh_cache-bd55e4b9-269f-4ca4-8139-86dface352b0" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1304.158167] env[63088]: DEBUG oslo_concurrency.lockutils [req-faf61559-e210-4983-94b5-4e071ac273d7 req-8f3ff297-7c9e-4a2a-a232-bb017e47926e service nova] Acquired lock "refresh_cache-bd55e4b9-269f-4ca4-8139-86dface352b0" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1304.158460] env[63088]: DEBUG nova.network.neutron [req-faf61559-e210-4983-94b5-4e071ac273d7 req-8f3ff297-7c9e-4a2a-a232-bb017e47926e service nova] [instance: bd55e4b9-269f-4ca4-8139-86dface352b0] Refreshing network info cache for port b40212cc-273b-4dbb-9363-91876eec40ad {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1304.269383] env[63088]: DEBUG oslo_vmware.rw_handles [None req-1160ba93-80cb-4c91-924c-c9b9da607442 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ab441d-04af-3a00-6c14-bcb020137cd5/disk-0.vmdk. {{(pid=63088) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1304.270398] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c55984f-a3e3-443e-9ceb-116cbc989be4 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1304.276738] env[63088]: DEBUG oslo_vmware.rw_handles [None req-1160ba93-80cb-4c91-924c-c9b9da607442 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ab441d-04af-3a00-6c14-bcb020137cd5/disk-0.vmdk is in state: ready. {{(pid=63088) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1304.276908] env[63088]: ERROR oslo_vmware.rw_handles [None req-1160ba93-80cb-4c91-924c-c9b9da607442 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ab441d-04af-3a00-6c14-bcb020137cd5/disk-0.vmdk due to incomplete transfer. [ 1304.277135] env[63088]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-725ff4b8-9605-4a27-9d25-ecfd5da2f5d2 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1304.283591] env[63088]: DEBUG oslo_vmware.rw_handles [None req-1160ba93-80cb-4c91-924c-c9b9da607442 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ab441d-04af-3a00-6c14-bcb020137cd5/disk-0.vmdk. {{(pid=63088) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1304.283812] env[63088]: DEBUG nova.virt.vmwareapi.images [None req-1160ba93-80cb-4c91-924c-c9b9da607442 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Uploaded image 5bb4b00f-be25-4207-8793-a8341056e087 to the Glance image server {{(pid=63088) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1304.285969] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-1160ba93-80cb-4c91-924c-c9b9da607442 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Destroying the VM {{(pid=63088) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1304.286213] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-98d2aad7-1c16-4667-874f-354cf0de54f2 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1304.291843] env[63088]: DEBUG oslo_vmware.api [None req-1160ba93-80cb-4c91-924c-c9b9da607442 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1304.291843] env[63088]: value = "task-1285654" [ 1304.291843] env[63088]: _type = "Task" [ 1304.291843] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1304.299474] env[63088]: DEBUG oslo_vmware.api [None req-1160ba93-80cb-4c91-924c-c9b9da607442 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285654, 'name': Destroy_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1304.429039] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285653, 'name': CreateVM_Task, 'duration_secs': 0.321431} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1304.429200] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bd55e4b9-269f-4ca4-8139-86dface352b0] Created VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1304.429822] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1304.429985] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1304.430321] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1304.430579] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3bd9cc0c-d2a6-46db-9f07-0750c8b9dc54 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1304.434941] env[63088]: DEBUG oslo_vmware.api [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Waiting for the task: (returnval){ [ 1304.434941] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]520c5db8-f145-51fb-77ea-453de5685c2f" [ 1304.434941] env[63088]: _type = "Task" [ 1304.434941] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1304.442433] env[63088]: DEBUG oslo_vmware.api [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]520c5db8-f145-51fb-77ea-453de5685c2f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1304.801355] env[63088]: DEBUG oslo_vmware.api [None req-1160ba93-80cb-4c91-924c-c9b9da607442 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285654, 'name': Destroy_Task, 'duration_secs': 0.286904} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1304.803455] env[63088]: INFO nova.virt.vmwareapi.vm_util [None req-1160ba93-80cb-4c91-924c-c9b9da607442 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Destroyed the VM [ 1304.803699] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-1160ba93-80cb-4c91-924c-c9b9da607442 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Deleting Snapshot of the VM instance {{(pid=63088) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1304.803952] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-7565f5f8-309c-41e8-a1fc-076ae28ed47f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1304.810166] env[63088]: DEBUG oslo_vmware.api [None req-1160ba93-80cb-4c91-924c-c9b9da607442 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1304.810166] env[63088]: value = "task-1285655" [ 1304.810166] env[63088]: _type = "Task" [ 1304.810166] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1304.817277] env[63088]: DEBUG oslo_vmware.api [None req-1160ba93-80cb-4c91-924c-c9b9da607442 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285655, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1304.838283] env[63088]: DEBUG nova.network.neutron [req-faf61559-e210-4983-94b5-4e071ac273d7 req-8f3ff297-7c9e-4a2a-a232-bb017e47926e service nova] [instance: bd55e4b9-269f-4ca4-8139-86dface352b0] Updated VIF entry in instance network info cache for port b40212cc-273b-4dbb-9363-91876eec40ad. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1304.838638] env[63088]: DEBUG nova.network.neutron [req-faf61559-e210-4983-94b5-4e071ac273d7 req-8f3ff297-7c9e-4a2a-a232-bb017e47926e service nova] [instance: bd55e4b9-269f-4ca4-8139-86dface352b0] Updating instance_info_cache with network_info: [{"id": "b40212cc-273b-4dbb-9363-91876eec40ad", "address": "fa:16:3e:36:62:93", "network": {"id": "ab40c372-54e5-49d6-a02a-279901071b67", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1152691158-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e4d8e091fab04519a6c430290830998c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb40212cc-27", "ovs_interfaceid": "b40212cc-273b-4dbb-9363-91876eec40ad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1304.945455] env[63088]: DEBUG oslo_vmware.api [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]520c5db8-f145-51fb-77ea-453de5685c2f, 'name': SearchDatastore_Task, 'duration_secs': 0.009672} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1304.945738] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1304.946033] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: bd55e4b9-269f-4ca4-8139-86dface352b0] Processing image 1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1304.946320] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1304.946481] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1304.946723] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1304.947028] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-706c43d3-c239-4c77-a0bd-ea3ddd894968 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1304.954628] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1304.954826] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1304.955545] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-41c3e7fb-c8f7-415e-88fd-477287fec1fd {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1304.960286] env[63088]: DEBUG oslo_vmware.api [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Waiting for the task: (returnval){ [ 1304.960286] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52b90c2a-dbab-bcd1-d106-e9f06aa0ef1d" [ 1304.960286] env[63088]: _type = "Task" [ 1304.960286] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1304.967162] env[63088]: DEBUG oslo_vmware.api [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52b90c2a-dbab-bcd1-d106-e9f06aa0ef1d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1305.320345] env[63088]: DEBUG oslo_vmware.api [None req-1160ba93-80cb-4c91-924c-c9b9da607442 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285655, 'name': RemoveSnapshot_Task, 'duration_secs': 0.370618} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1305.320697] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-1160ba93-80cb-4c91-924c-c9b9da607442 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Deleted Snapshot of the VM instance {{(pid=63088) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1305.320897] env[63088]: DEBUG nova.compute.manager [None req-1160ba93-80cb-4c91-924c-c9b9da607442 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1305.321647] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1287f41b-b88f-475c-8238-ff2391765226 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1305.341970] env[63088]: DEBUG oslo_concurrency.lockutils [req-faf61559-e210-4983-94b5-4e071ac273d7 req-8f3ff297-7c9e-4a2a-a232-bb017e47926e service nova] Releasing lock "refresh_cache-bd55e4b9-269f-4ca4-8139-86dface352b0" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1305.470324] env[63088]: DEBUG oslo_vmware.api [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52b90c2a-dbab-bcd1-d106-e9f06aa0ef1d, 'name': SearchDatastore_Task, 'duration_secs': 0.007927} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1305.471043] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f330be12-0813-455a-9aa0-9acc9f2a8481 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1305.475739] env[63088]: DEBUG oslo_vmware.api [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Waiting for the task: (returnval){ [ 1305.475739] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]525444b4-ce06-3274-f4be-e26bb1bcfbda" [ 1305.475739] env[63088]: _type = "Task" [ 1305.475739] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1305.483655] env[63088]: DEBUG oslo_vmware.api [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]525444b4-ce06-3274-f4be-e26bb1bcfbda, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1305.833105] env[63088]: INFO nova.compute.manager [None req-1160ba93-80cb-4c91-924c-c9b9da607442 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Shelve offloading [ 1305.834721] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-1160ba93-80cb-4c91-924c-c9b9da607442 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1305.834971] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-092254f7-0f3f-4377-a6be-074cbdec9143 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1305.841759] env[63088]: DEBUG oslo_vmware.api [None req-1160ba93-80cb-4c91-924c-c9b9da607442 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1305.841759] env[63088]: value = "task-1285656" [ 1305.841759] env[63088]: _type = "Task" [ 1305.841759] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1305.849299] env[63088]: DEBUG oslo_vmware.api [None req-1160ba93-80cb-4c91-924c-c9b9da607442 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285656, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1305.986027] env[63088]: DEBUG oslo_vmware.api [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]525444b4-ce06-3274-f4be-e26bb1bcfbda, 'name': SearchDatastore_Task, 'duration_secs': 0.009361} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1305.986366] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1305.986644] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore2] bd55e4b9-269f-4ca4-8139-86dface352b0/bd55e4b9-269f-4ca4-8139-86dface352b0.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1305.986900] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-16a9769a-7a2c-4736-a113-a436ce192145 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1305.992396] env[63088]: DEBUG oslo_vmware.api [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Waiting for the task: (returnval){ [ 1305.992396] env[63088]: value = "task-1285657" [ 1305.992396] env[63088]: _type = "Task" [ 1305.992396] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1305.999464] env[63088]: DEBUG oslo_vmware.api [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285657, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1306.353358] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-1160ba93-80cb-4c91-924c-c9b9da607442 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] VM already powered off {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1306.353792] env[63088]: DEBUG nova.compute.manager [None req-1160ba93-80cb-4c91-924c-c9b9da607442 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1306.354448] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-624d7dd0-4b36-4c52-a687-24be1f8c7a8b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1306.360861] env[63088]: DEBUG oslo_concurrency.lockutils [None req-1160ba93-80cb-4c91-924c-c9b9da607442 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquiring lock "refresh_cache-3d23868c-b0fd-470e-970c-c9935739b7b7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1306.361092] env[63088]: DEBUG oslo_concurrency.lockutils [None req-1160ba93-80cb-4c91-924c-c9b9da607442 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquired lock "refresh_cache-3d23868c-b0fd-470e-970c-c9935739b7b7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1306.361244] env[63088]: DEBUG nova.network.neutron [None req-1160ba93-80cb-4c91-924c-c9b9da607442 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1306.503080] env[63088]: DEBUG oslo_vmware.api [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285657, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.504779} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1306.503449] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore2] bd55e4b9-269f-4ca4-8139-86dface352b0/bd55e4b9-269f-4ca4-8139-86dface352b0.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1306.503792] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: bd55e4b9-269f-4ca4-8139-86dface352b0] Extending root virtual disk to 1048576 {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1306.504117] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-952219d5-64a2-4b34-960c-58dda889dd57 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1306.510300] env[63088]: DEBUG oslo_vmware.api [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Waiting for the task: (returnval){ [ 1306.510300] env[63088]: value = "task-1285658" [ 1306.510300] env[63088]: _type = "Task" [ 1306.510300] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1306.517131] env[63088]: DEBUG oslo_vmware.api [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285658, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1307.019831] env[63088]: DEBUG oslo_vmware.api [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285658, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.053182} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1307.020104] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: bd55e4b9-269f-4ca4-8139-86dface352b0] Extended root virtual disk {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1307.020851] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-052f9d9a-3f94-4159-8ac1-c709c9560e7e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1307.041928] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: bd55e4b9-269f-4ca4-8139-86dface352b0] Reconfiguring VM instance instance-0000006e to attach disk [datastore2] bd55e4b9-269f-4ca4-8139-86dface352b0/bd55e4b9-269f-4ca4-8139-86dface352b0.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1307.044099] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e44c8582-1a31-4594-89a7-47976edf9872 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1307.062635] env[63088]: DEBUG oslo_vmware.api [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Waiting for the task: (returnval){ [ 1307.062635] env[63088]: value = "task-1285659" [ 1307.062635] env[63088]: _type = "Task" [ 1307.062635] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1307.070082] env[63088]: DEBUG oslo_vmware.api [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285659, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1307.090594] env[63088]: DEBUG nova.network.neutron [None req-1160ba93-80cb-4c91-924c-c9b9da607442 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Updating instance_info_cache with network_info: [{"id": "e7ade0ec-6647-4782-8915-4edcb26c64e1", "address": "fa:16:3e:43:c0:3b", "network": {"id": "711d3d9a-86db-4aa1-87b5-481f1812dfa4", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-59376186-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.225", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a7898ddafe0d41038a6ae4277f2c2f48", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e41070eb-3ac1-4ca9-a3d0-fd65893a97de", "external-id": "nsx-vlan-transportzone-596", "segmentation_id": 596, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape7ade0ec-66", "ovs_interfaceid": "e7ade0ec-6647-4782-8915-4edcb26c64e1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1307.572643] env[63088]: DEBUG oslo_vmware.api [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285659, 'name': ReconfigVM_Task, 'duration_secs': 0.32188} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1307.573018] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: bd55e4b9-269f-4ca4-8139-86dface352b0] Reconfigured VM instance instance-0000006e to attach disk [datastore2] bd55e4b9-269f-4ca4-8139-86dface352b0/bd55e4b9-269f-4ca4-8139-86dface352b0.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1307.573573] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-89670168-f8e7-4e99-8846-8bf7a04a4f28 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1307.579791] env[63088]: DEBUG oslo_vmware.api [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Waiting for the task: (returnval){ [ 1307.579791] env[63088]: value = "task-1285660" [ 1307.579791] env[63088]: _type = "Task" [ 1307.579791] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1307.587956] env[63088]: DEBUG oslo_vmware.api [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285660, 'name': Rename_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1307.593453] env[63088]: DEBUG oslo_concurrency.lockutils [None req-1160ba93-80cb-4c91-924c-c9b9da607442 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Releasing lock "refresh_cache-3d23868c-b0fd-470e-970c-c9935739b7b7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1307.800388] env[63088]: DEBUG nova.compute.manager [req-ea46509c-12a6-4938-b0d2-dc3cf52251d7 req-33a97627-cb7e-4e5f-b60e-5b11e677c94d service nova] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Received event network-vif-unplugged-e7ade0ec-6647-4782-8915-4edcb26c64e1 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1307.800566] env[63088]: DEBUG oslo_concurrency.lockutils [req-ea46509c-12a6-4938-b0d2-dc3cf52251d7 req-33a97627-cb7e-4e5f-b60e-5b11e677c94d service nova] Acquiring lock "3d23868c-b0fd-470e-970c-c9935739b7b7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1307.800777] env[63088]: DEBUG oslo_concurrency.lockutils [req-ea46509c-12a6-4938-b0d2-dc3cf52251d7 req-33a97627-cb7e-4e5f-b60e-5b11e677c94d service nova] Lock "3d23868c-b0fd-470e-970c-c9935739b7b7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1307.800953] env[63088]: DEBUG oslo_concurrency.lockutils [req-ea46509c-12a6-4938-b0d2-dc3cf52251d7 req-33a97627-cb7e-4e5f-b60e-5b11e677c94d service nova] Lock "3d23868c-b0fd-470e-970c-c9935739b7b7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1307.801144] env[63088]: DEBUG nova.compute.manager [req-ea46509c-12a6-4938-b0d2-dc3cf52251d7 req-33a97627-cb7e-4e5f-b60e-5b11e677c94d service nova] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] No waiting events found dispatching network-vif-unplugged-e7ade0ec-6647-4782-8915-4edcb26c64e1 {{(pid=63088) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1307.801317] env[63088]: WARNING nova.compute.manager [req-ea46509c-12a6-4938-b0d2-dc3cf52251d7 req-33a97627-cb7e-4e5f-b60e-5b11e677c94d service nova] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Received unexpected event network-vif-unplugged-e7ade0ec-6647-4782-8915-4edcb26c64e1 for instance with vm_state shelved and task_state shelving_offloading. [ 1307.887571] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-1160ba93-80cb-4c91-924c-c9b9da607442 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1307.888501] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63719910-aae7-466f-b57b-1de4fb623379 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1307.896031] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-1160ba93-80cb-4c91-924c-c9b9da607442 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Unregistering the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1307.896306] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ac7b5b88-a8d7-4f06-a70d-0b8ee32206c8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1307.959065] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-1160ba93-80cb-4c91-924c-c9b9da607442 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Unregistered the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1307.959314] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-1160ba93-80cb-4c91-924c-c9b9da607442 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Deleting contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1307.959545] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-1160ba93-80cb-4c91-924c-c9b9da607442 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Deleting the datastore file [datastore1] 3d23868c-b0fd-470e-970c-c9935739b7b7 {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1307.959842] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4cc15579-4117-44be-abe2-6cb196d2f081 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1307.966497] env[63088]: DEBUG oslo_vmware.api [None req-1160ba93-80cb-4c91-924c-c9b9da607442 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1307.966497] env[63088]: value = "task-1285662" [ 1307.966497] env[63088]: _type = "Task" [ 1307.966497] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1307.973679] env[63088]: DEBUG oslo_vmware.api [None req-1160ba93-80cb-4c91-924c-c9b9da607442 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285662, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1308.090175] env[63088]: DEBUG oslo_vmware.api [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285660, 'name': Rename_Task, 'duration_secs': 0.130524} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1308.090477] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: bd55e4b9-269f-4ca4-8139-86dface352b0] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1308.090737] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-116123fd-665c-4056-adec-2de84a8e3ad2 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1308.097252] env[63088]: DEBUG oslo_vmware.api [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Waiting for the task: (returnval){ [ 1308.097252] env[63088]: value = "task-1285663" [ 1308.097252] env[63088]: _type = "Task" [ 1308.097252] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1308.108544] env[63088]: DEBUG oslo_vmware.api [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285663, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1308.476393] env[63088]: DEBUG oslo_vmware.api [None req-1160ba93-80cb-4c91-924c-c9b9da607442 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285662, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.135048} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1308.476674] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-1160ba93-80cb-4c91-924c-c9b9da607442 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1308.476828] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-1160ba93-80cb-4c91-924c-c9b9da607442 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Deleted contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1308.477011] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-1160ba93-80cb-4c91-924c-c9b9da607442 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1308.501877] env[63088]: INFO nova.scheduler.client.report [None req-1160ba93-80cb-4c91-924c-c9b9da607442 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Deleted allocations for instance 3d23868c-b0fd-470e-970c-c9935739b7b7 [ 1308.607351] env[63088]: DEBUG oslo_vmware.api [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285663, 'name': PowerOnVM_Task, 'duration_secs': 0.40302} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1308.607670] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: bd55e4b9-269f-4ca4-8139-86dface352b0] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1308.607873] env[63088]: INFO nova.compute.manager [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: bd55e4b9-269f-4ca4-8139-86dface352b0] Took 6.69 seconds to spawn the instance on the hypervisor. [ 1308.608065] env[63088]: DEBUG nova.compute.manager [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: bd55e4b9-269f-4ca4-8139-86dface352b0] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1308.608819] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d96622b0-c734-4ddf-b682-dbdda8c169e6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1309.007118] env[63088]: DEBUG oslo_concurrency.lockutils [None req-1160ba93-80cb-4c91-924c-c9b9da607442 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1309.007372] env[63088]: DEBUG oslo_concurrency.lockutils [None req-1160ba93-80cb-4c91-924c-c9b9da607442 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1309.007627] env[63088]: DEBUG nova.objects.instance [None req-1160ba93-80cb-4c91-924c-c9b9da607442 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lazy-loading 'resources' on Instance uuid 3d23868c-b0fd-470e-970c-c9935739b7b7 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1309.124800] env[63088]: INFO nova.compute.manager [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: bd55e4b9-269f-4ca4-8139-86dface352b0] Took 11.38 seconds to build instance. [ 1309.510790] env[63088]: DEBUG nova.objects.instance [None req-1160ba93-80cb-4c91-924c-c9b9da607442 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lazy-loading 'numa_topology' on Instance uuid 3d23868c-b0fd-470e-970c-c9935739b7b7 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1309.626939] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0d406dc0-c42b-4119-90cb-157a35c988d5 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Lock "bd55e4b9-269f-4ca4-8139-86dface352b0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.888s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1309.792496] env[63088]: DEBUG nova.compute.manager [req-61233fe2-8810-4172-a311-dc35b4655ef1 req-bb279394-3a60-49e6-94c7-e3ea89bebd53 service nova] [instance: bd55e4b9-269f-4ca4-8139-86dface352b0] Received event network-changed-b40212cc-273b-4dbb-9363-91876eec40ad {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1309.792694] env[63088]: DEBUG nova.compute.manager [req-61233fe2-8810-4172-a311-dc35b4655ef1 req-bb279394-3a60-49e6-94c7-e3ea89bebd53 service nova] [instance: bd55e4b9-269f-4ca4-8139-86dface352b0] Refreshing instance network info cache due to event network-changed-b40212cc-273b-4dbb-9363-91876eec40ad. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1309.792916] env[63088]: DEBUG oslo_concurrency.lockutils [req-61233fe2-8810-4172-a311-dc35b4655ef1 req-bb279394-3a60-49e6-94c7-e3ea89bebd53 service nova] Acquiring lock "refresh_cache-bd55e4b9-269f-4ca4-8139-86dface352b0" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1309.793096] env[63088]: DEBUG oslo_concurrency.lockutils [req-61233fe2-8810-4172-a311-dc35b4655ef1 req-bb279394-3a60-49e6-94c7-e3ea89bebd53 service nova] Acquired lock "refresh_cache-bd55e4b9-269f-4ca4-8139-86dface352b0" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1309.793297] env[63088]: DEBUG nova.network.neutron [req-61233fe2-8810-4172-a311-dc35b4655ef1 req-bb279394-3a60-49e6-94c7-e3ea89bebd53 service nova] [instance: bd55e4b9-269f-4ca4-8139-86dface352b0] Refreshing network info cache for port b40212cc-273b-4dbb-9363-91876eec40ad {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1309.822635] env[63088]: DEBUG nova.compute.manager [req-0e928774-a264-46ad-a3b1-cb326c0fbe7e req-e465a4c2-d004-40e1-ac1a-5e0e9505737a service nova] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Received event network-changed-e7ade0ec-6647-4782-8915-4edcb26c64e1 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1309.822975] env[63088]: DEBUG nova.compute.manager [req-0e928774-a264-46ad-a3b1-cb326c0fbe7e req-e465a4c2-d004-40e1-ac1a-5e0e9505737a service nova] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Refreshing instance network info cache due to event network-changed-e7ade0ec-6647-4782-8915-4edcb26c64e1. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1309.823041] env[63088]: DEBUG oslo_concurrency.lockutils [req-0e928774-a264-46ad-a3b1-cb326c0fbe7e req-e465a4c2-d004-40e1-ac1a-5e0e9505737a service nova] Acquiring lock "refresh_cache-3d23868c-b0fd-470e-970c-c9935739b7b7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1309.823198] env[63088]: DEBUG oslo_concurrency.lockutils [req-0e928774-a264-46ad-a3b1-cb326c0fbe7e req-e465a4c2-d004-40e1-ac1a-5e0e9505737a service nova] Acquired lock "refresh_cache-3d23868c-b0fd-470e-970c-c9935739b7b7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1309.823353] env[63088]: DEBUG nova.network.neutron [req-0e928774-a264-46ad-a3b1-cb326c0fbe7e req-e465a4c2-d004-40e1-ac1a-5e0e9505737a service nova] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Refreshing network info cache for port e7ade0ec-6647-4782-8915-4edcb26c64e1 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1310.013142] env[63088]: DEBUG nova.objects.base [None req-1160ba93-80cb-4c91-924c-c9b9da607442 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Object Instance<3d23868c-b0fd-470e-970c-c9935739b7b7> lazy-loaded attributes: resources,numa_topology {{(pid=63088) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1310.062857] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdd8d32e-744d-4345-8184-3c2aa636abce {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1310.070299] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d5e18f7-a651-4057-8897-74e2bbe0c2f7 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1310.100494] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db0849b1-f49c-4a6c-bec7-acfe961c15f0 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1310.107510] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2569c520-59bb-48d6-8285-8e09d9569ba9 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1310.120435] env[63088]: DEBUG nova.compute.provider_tree [None req-1160ba93-80cb-4c91-924c-c9b9da607442 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1310.623264] env[63088]: DEBUG nova.scheduler.client.report [None req-1160ba93-80cb-4c91-924c-c9b9da607442 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1310.626890] env[63088]: DEBUG nova.network.neutron [req-61233fe2-8810-4172-a311-dc35b4655ef1 req-bb279394-3a60-49e6-94c7-e3ea89bebd53 service nova] [instance: bd55e4b9-269f-4ca4-8139-86dface352b0] Updated VIF entry in instance network info cache for port b40212cc-273b-4dbb-9363-91876eec40ad. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1310.627281] env[63088]: DEBUG nova.network.neutron [req-61233fe2-8810-4172-a311-dc35b4655ef1 req-bb279394-3a60-49e6-94c7-e3ea89bebd53 service nova] [instance: bd55e4b9-269f-4ca4-8139-86dface352b0] Updating instance_info_cache with network_info: [{"id": "b40212cc-273b-4dbb-9363-91876eec40ad", "address": "fa:16:3e:36:62:93", "network": {"id": "ab40c372-54e5-49d6-a02a-279901071b67", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1152691158-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.205", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e4d8e091fab04519a6c430290830998c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb40212cc-27", "ovs_interfaceid": "b40212cc-273b-4dbb-9363-91876eec40ad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1310.650713] env[63088]: DEBUG nova.network.neutron [req-0e928774-a264-46ad-a3b1-cb326c0fbe7e req-e465a4c2-d004-40e1-ac1a-5e0e9505737a service nova] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Updated VIF entry in instance network info cache for port e7ade0ec-6647-4782-8915-4edcb26c64e1. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1310.651063] env[63088]: DEBUG nova.network.neutron [req-0e928774-a264-46ad-a3b1-cb326c0fbe7e req-e465a4c2-d004-40e1-ac1a-5e0e9505737a service nova] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Updating instance_info_cache with network_info: [{"id": "e7ade0ec-6647-4782-8915-4edcb26c64e1", "address": "fa:16:3e:43:c0:3b", "network": {"id": "711d3d9a-86db-4aa1-87b5-481f1812dfa4", "bridge": null, "label": "tempest-ServerActionsTestOtherB-59376186-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.225", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a7898ddafe0d41038a6ae4277f2c2f48", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tape7ade0ec-66", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1311.129674] env[63088]: DEBUG oslo_concurrency.lockutils [None req-1160ba93-80cb-4c91-924c-c9b9da607442 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.122s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1311.132296] env[63088]: DEBUG oslo_concurrency.lockutils [req-61233fe2-8810-4172-a311-dc35b4655ef1 req-bb279394-3a60-49e6-94c7-e3ea89bebd53 service nova] Releasing lock "refresh_cache-bd55e4b9-269f-4ca4-8139-86dface352b0" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1311.153535] env[63088]: DEBUG oslo_concurrency.lockutils [req-0e928774-a264-46ad-a3b1-cb326c0fbe7e req-e465a4c2-d004-40e1-ac1a-5e0e9505737a service nova] Releasing lock "refresh_cache-3d23868c-b0fd-470e-970c-c9935739b7b7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1311.171526] env[63088]: DEBUG oslo_concurrency.lockutils [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquiring lock "3d23868c-b0fd-470e-970c-c9935739b7b7" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1311.638180] env[63088]: DEBUG oslo_concurrency.lockutils [None req-1160ba93-80cb-4c91-924c-c9b9da607442 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lock "3d23868c-b0fd-470e-970c-c9935739b7b7" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 18.233s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1311.639019] env[63088]: DEBUG oslo_concurrency.lockutils [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lock "3d23868c-b0fd-470e-970c-c9935739b7b7" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 0.468s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1311.639215] env[63088]: INFO nova.compute.manager [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Unshelving [ 1312.663089] env[63088]: DEBUG oslo_concurrency.lockutils [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1312.663413] env[63088]: DEBUG oslo_concurrency.lockutils [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1312.663635] env[63088]: DEBUG nova.objects.instance [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lazy-loading 'pci_requests' on Instance uuid 3d23868c-b0fd-470e-970c-c9935739b7b7 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1313.167542] env[63088]: DEBUG nova.objects.instance [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lazy-loading 'numa_topology' on Instance uuid 3d23868c-b0fd-470e-970c-c9935739b7b7 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1313.670697] env[63088]: INFO nova.compute.claims [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1314.724844] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bea43190-2088-4cf4-a952-b8784f936ec3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1314.732147] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe194860-9285-4203-b8cf-2c2ecd0e915e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1314.761787] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-591e9614-d8b3-41c2-bc04-5f0887db6ca1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1314.768597] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-025e4bfc-85a0-4f79-9be6-1d20a4be0f29 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1314.781173] env[63088]: DEBUG nova.compute.provider_tree [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1315.284313] env[63088]: DEBUG nova.scheduler.client.report [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1315.789331] env[63088]: DEBUG oslo_concurrency.lockutils [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.126s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1315.817647] env[63088]: INFO nova.network.neutron [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Updating port e7ade0ec-6647-4782-8915-4edcb26c64e1 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1317.220671] env[63088]: DEBUG nova.compute.manager [req-223979f8-cca3-47f6-a620-d5cf8d72371b req-2376f374-28ac-4f62-a6c7-f8dfe90eea84 service nova] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Received event network-vif-plugged-e7ade0ec-6647-4782-8915-4edcb26c64e1 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1317.220916] env[63088]: DEBUG oslo_concurrency.lockutils [req-223979f8-cca3-47f6-a620-d5cf8d72371b req-2376f374-28ac-4f62-a6c7-f8dfe90eea84 service nova] Acquiring lock "3d23868c-b0fd-470e-970c-c9935739b7b7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1317.221147] env[63088]: DEBUG oslo_concurrency.lockutils [req-223979f8-cca3-47f6-a620-d5cf8d72371b req-2376f374-28ac-4f62-a6c7-f8dfe90eea84 service nova] Lock "3d23868c-b0fd-470e-970c-c9935739b7b7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1317.221331] env[63088]: DEBUG oslo_concurrency.lockutils [req-223979f8-cca3-47f6-a620-d5cf8d72371b req-2376f374-28ac-4f62-a6c7-f8dfe90eea84 service nova] Lock "3d23868c-b0fd-470e-970c-c9935739b7b7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1317.221617] env[63088]: DEBUG nova.compute.manager [req-223979f8-cca3-47f6-a620-d5cf8d72371b req-2376f374-28ac-4f62-a6c7-f8dfe90eea84 service nova] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] No waiting events found dispatching network-vif-plugged-e7ade0ec-6647-4782-8915-4edcb26c64e1 {{(pid=63088) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1317.221862] env[63088]: WARNING nova.compute.manager [req-223979f8-cca3-47f6-a620-d5cf8d72371b req-2376f374-28ac-4f62-a6c7-f8dfe90eea84 service nova] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Received unexpected event network-vif-plugged-e7ade0ec-6647-4782-8915-4edcb26c64e1 for instance with vm_state shelved_offloaded and task_state spawning. [ 1317.308068] env[63088]: DEBUG oslo_concurrency.lockutils [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquiring lock "refresh_cache-3d23868c-b0fd-470e-970c-c9935739b7b7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1317.308278] env[63088]: DEBUG oslo_concurrency.lockutils [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquired lock "refresh_cache-3d23868c-b0fd-470e-970c-c9935739b7b7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1317.308502] env[63088]: DEBUG nova.network.neutron [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1318.010876] env[63088]: DEBUG nova.network.neutron [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Updating instance_info_cache with network_info: [{"id": "e7ade0ec-6647-4782-8915-4edcb26c64e1", "address": "fa:16:3e:43:c0:3b", "network": {"id": "711d3d9a-86db-4aa1-87b5-481f1812dfa4", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-59376186-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.225", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a7898ddafe0d41038a6ae4277f2c2f48", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e41070eb-3ac1-4ca9-a3d0-fd65893a97de", "external-id": "nsx-vlan-transportzone-596", "segmentation_id": 596, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape7ade0ec-66", "ovs_interfaceid": "e7ade0ec-6647-4782-8915-4edcb26c64e1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1318.514518] env[63088]: DEBUG oslo_concurrency.lockutils [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Releasing lock "refresh_cache-3d23868c-b0fd-470e-970c-c9935739b7b7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1318.540378] env[63088]: DEBUG nova.virt.hardware [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='941ef4cb7db3a9fa063204afcfb0bd2c',container_format='bare',created_at=2024-10-15T12:26:42Z,direct_url=,disk_format='vmdk',id=5bb4b00f-be25-4207-8793-a8341056e087,min_disk=1,min_ram=0,name='tempest-ServerActionsTestOtherB-server-1365142169-shelved',owner='a7898ddafe0d41038a6ae4277f2c2f48',properties=ImageMetaProps,protected=,size=31669760,status='active',tags=,updated_at=2024-10-15T12:26:54Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1318.540658] env[63088]: DEBUG nova.virt.hardware [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1318.540822] env[63088]: DEBUG nova.virt.hardware [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1318.541016] env[63088]: DEBUG nova.virt.hardware [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1318.541174] env[63088]: DEBUG nova.virt.hardware [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1318.541350] env[63088]: DEBUG nova.virt.hardware [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1318.541532] env[63088]: DEBUG nova.virt.hardware [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1318.541696] env[63088]: DEBUG nova.virt.hardware [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1318.541877] env[63088]: DEBUG nova.virt.hardware [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1318.542043] env[63088]: DEBUG nova.virt.hardware [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1318.542236] env[63088]: DEBUG nova.virt.hardware [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1318.543109] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a08df398-fd4e-4f59-b981-0ce0900adb79 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1318.550758] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88387968-fc33-47ac-bc01-1615abc156ed {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1318.563662] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:43:c0:3b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e41070eb-3ac1-4ca9-a3d0-fd65893a97de', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e7ade0ec-6647-4782-8915-4edcb26c64e1', 'vif_model': 'vmxnet3'}] {{(pid=63088) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1318.570998] env[63088]: DEBUG oslo.service.loopingcall [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1318.571270] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Creating VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1318.571499] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1f565844-a53a-48ea-9895-29c17699bcd6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1318.589583] env[63088]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1318.589583] env[63088]: value = "task-1285664" [ 1318.589583] env[63088]: _type = "Task" [ 1318.589583] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1318.596355] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285664, 'name': CreateVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1319.099548] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285664, 'name': CreateVM_Task, 'duration_secs': 0.28213} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1319.099704] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Created VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1319.100343] env[63088]: DEBUG oslo_concurrency.lockutils [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5bb4b00f-be25-4207-8793-a8341056e087" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1319.100525] env[63088]: DEBUG oslo_concurrency.lockutils [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5bb4b00f-be25-4207-8793-a8341056e087" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1319.100903] env[63088]: DEBUG oslo_concurrency.lockutils [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/5bb4b00f-be25-4207-8793-a8341056e087" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1319.101171] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c8a2d890-9a69-4bb4-985f-425aa7cd41fd {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1319.105226] env[63088]: DEBUG oslo_vmware.api [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1319.105226] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]5224b684-7714-8b4c-ead9-756733d5c10b" [ 1319.105226] env[63088]: _type = "Task" [ 1319.105226] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1319.113155] env[63088]: DEBUG oslo_vmware.api [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]5224b684-7714-8b4c-ead9-756733d5c10b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1319.247700] env[63088]: DEBUG nova.compute.manager [req-05ded2b1-819f-4d32-bc90-feb100d84991 req-7dc4d605-707e-4391-856d-1869edfbf480 service nova] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Received event network-changed-e7ade0ec-6647-4782-8915-4edcb26c64e1 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1319.247908] env[63088]: DEBUG nova.compute.manager [req-05ded2b1-819f-4d32-bc90-feb100d84991 req-7dc4d605-707e-4391-856d-1869edfbf480 service nova] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Refreshing instance network info cache due to event network-changed-e7ade0ec-6647-4782-8915-4edcb26c64e1. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1319.248168] env[63088]: DEBUG oslo_concurrency.lockutils [req-05ded2b1-819f-4d32-bc90-feb100d84991 req-7dc4d605-707e-4391-856d-1869edfbf480 service nova] Acquiring lock "refresh_cache-3d23868c-b0fd-470e-970c-c9935739b7b7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1319.248320] env[63088]: DEBUG oslo_concurrency.lockutils [req-05ded2b1-819f-4d32-bc90-feb100d84991 req-7dc4d605-707e-4391-856d-1869edfbf480 service nova] Acquired lock "refresh_cache-3d23868c-b0fd-470e-970c-c9935739b7b7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1319.248487] env[63088]: DEBUG nova.network.neutron [req-05ded2b1-819f-4d32-bc90-feb100d84991 req-7dc4d605-707e-4391-856d-1869edfbf480 service nova] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Refreshing network info cache for port e7ade0ec-6647-4782-8915-4edcb26c64e1 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1319.615449] env[63088]: DEBUG oslo_concurrency.lockutils [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5bb4b00f-be25-4207-8793-a8341056e087" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1319.615835] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Processing image 5bb4b00f-be25-4207-8793-a8341056e087 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1319.615967] env[63088]: DEBUG oslo_concurrency.lockutils [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5bb4b00f-be25-4207-8793-a8341056e087/5bb4b00f-be25-4207-8793-a8341056e087.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1319.616135] env[63088]: DEBUG oslo_concurrency.lockutils [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5bb4b00f-be25-4207-8793-a8341056e087/5bb4b00f-be25-4207-8793-a8341056e087.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1319.616328] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1319.616606] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f32ebfb5-68b4-4544-b549-44221f081077 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1319.624521] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1319.624683] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1319.625383] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dca91afd-aa73-400c-857c-3761336e966e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1319.629856] env[63088]: DEBUG oslo_vmware.api [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1319.629856] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]5276de26-8a6b-f4e2-09fe-bff3016d98c0" [ 1319.629856] env[63088]: _type = "Task" [ 1319.629856] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1319.636692] env[63088]: DEBUG oslo_vmware.api [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]5276de26-8a6b-f4e2-09fe-bff3016d98c0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1319.967315] env[63088]: DEBUG nova.network.neutron [req-05ded2b1-819f-4d32-bc90-feb100d84991 req-7dc4d605-707e-4391-856d-1869edfbf480 service nova] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Updated VIF entry in instance network info cache for port e7ade0ec-6647-4782-8915-4edcb26c64e1. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1319.968048] env[63088]: DEBUG nova.network.neutron [req-05ded2b1-819f-4d32-bc90-feb100d84991 req-7dc4d605-707e-4391-856d-1869edfbf480 service nova] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Updating instance_info_cache with network_info: [{"id": "e7ade0ec-6647-4782-8915-4edcb26c64e1", "address": "fa:16:3e:43:c0:3b", "network": {"id": "711d3d9a-86db-4aa1-87b5-481f1812dfa4", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-59376186-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.225", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a7898ddafe0d41038a6ae4277f2c2f48", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e41070eb-3ac1-4ca9-a3d0-fd65893a97de", "external-id": "nsx-vlan-transportzone-596", "segmentation_id": 596, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape7ade0ec-66", "ovs_interfaceid": "e7ade0ec-6647-4782-8915-4edcb26c64e1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1320.139510] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Preparing fetch location {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1320.139870] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Fetch image to [datastore1] OSTACK_IMG_cc617ad8-1565-4ed3-87d9-799f96b6381a/OSTACK_IMG_cc617ad8-1565-4ed3-87d9-799f96b6381a.vmdk {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1320.140122] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Downloading stream optimized image 5bb4b00f-be25-4207-8793-a8341056e087 to [datastore1] OSTACK_IMG_cc617ad8-1565-4ed3-87d9-799f96b6381a/OSTACK_IMG_cc617ad8-1565-4ed3-87d9-799f96b6381a.vmdk on the data store datastore1 as vApp {{(pid=63088) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1320.140340] env[63088]: DEBUG nova.virt.vmwareapi.images [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Downloading image file data 5bb4b00f-be25-4207-8793-a8341056e087 to the ESX as VM named 'OSTACK_IMG_cc617ad8-1565-4ed3-87d9-799f96b6381a' {{(pid=63088) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1320.204544] env[63088]: DEBUG oslo_vmware.rw_handles [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1320.204544] env[63088]: value = "resgroup-9" [ 1320.204544] env[63088]: _type = "ResourcePool" [ 1320.204544] env[63088]: }. {{(pid=63088) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1320.204897] env[63088]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-e365c317-2529-4338-996b-2601015122b0 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1320.224786] env[63088]: DEBUG oslo_vmware.rw_handles [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lease: (returnval){ [ 1320.224786] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]528be36b-1d53-0dbc-bc79-a0629992e430" [ 1320.224786] env[63088]: _type = "HttpNfcLease" [ 1320.224786] env[63088]: } obtained for vApp import into resource pool (val){ [ 1320.224786] env[63088]: value = "resgroup-9" [ 1320.224786] env[63088]: _type = "ResourcePool" [ 1320.224786] env[63088]: }. {{(pid=63088) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1320.225092] env[63088]: DEBUG oslo_vmware.api [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the lease: (returnval){ [ 1320.225092] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]528be36b-1d53-0dbc-bc79-a0629992e430" [ 1320.225092] env[63088]: _type = "HttpNfcLease" [ 1320.225092] env[63088]: } to be ready. {{(pid=63088) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1320.230937] env[63088]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1320.230937] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]528be36b-1d53-0dbc-bc79-a0629992e430" [ 1320.230937] env[63088]: _type = "HttpNfcLease" [ 1320.230937] env[63088]: } is initializing. {{(pid=63088) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1320.470978] env[63088]: DEBUG oslo_concurrency.lockutils [req-05ded2b1-819f-4d32-bc90-feb100d84991 req-7dc4d605-707e-4391-856d-1869edfbf480 service nova] Releasing lock "refresh_cache-3d23868c-b0fd-470e-970c-c9935739b7b7" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1320.733423] env[63088]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1320.733423] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]528be36b-1d53-0dbc-bc79-a0629992e430" [ 1320.733423] env[63088]: _type = "HttpNfcLease" [ 1320.733423] env[63088]: } is ready. {{(pid=63088) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1320.733900] env[63088]: DEBUG oslo_vmware.rw_handles [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1320.733900] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]528be36b-1d53-0dbc-bc79-a0629992e430" [ 1320.733900] env[63088]: _type = "HttpNfcLease" [ 1320.733900] env[63088]: }. {{(pid=63088) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1320.734381] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2fd5f09-5a74-4268-9dd1-da79fba0879f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1320.741325] env[63088]: DEBUG oslo_vmware.rw_handles [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52397e25-5d6b-f258-b3ae-d2a4567fd3c8/disk-0.vmdk from lease info. {{(pid=63088) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1320.741508] env[63088]: DEBUG oslo_vmware.rw_handles [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Creating HTTP connection to write to file with size = 31669760 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52397e25-5d6b-f258-b3ae-d2a4567fd3c8/disk-0.vmdk. {{(pid=63088) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1320.804996] env[63088]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-1a39c5ed-9cc6-4e95-84d3-feca7a8ba59f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1321.995982] env[63088]: DEBUG oslo_vmware.rw_handles [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Completed reading data from the image iterator. {{(pid=63088) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1321.996433] env[63088]: DEBUG oslo_vmware.rw_handles [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52397e25-5d6b-f258-b3ae-d2a4567fd3c8/disk-0.vmdk. {{(pid=63088) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1321.997395] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d03c5ff8-16c9-4d68-b6c7-3b9f82a1aba1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1322.004492] env[63088]: DEBUG oslo_vmware.rw_handles [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52397e25-5d6b-f258-b3ae-d2a4567fd3c8/disk-0.vmdk is in state: ready. {{(pid=63088) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1322.004687] env[63088]: DEBUG oslo_vmware.rw_handles [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52397e25-5d6b-f258-b3ae-d2a4567fd3c8/disk-0.vmdk. {{(pid=63088) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1322.004968] env[63088]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-a75fd9c3-b511-45b4-9de5-f5f17f5aacf3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1322.183961] env[63088]: DEBUG oslo_vmware.rw_handles [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52397e25-5d6b-f258-b3ae-d2a4567fd3c8/disk-0.vmdk. {{(pid=63088) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1322.184284] env[63088]: INFO nova.virt.vmwareapi.images [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Downloaded image file data 5bb4b00f-be25-4207-8793-a8341056e087 [ 1322.185076] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30f981cc-dc64-4ccd-9c6e-4a23255b536d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1322.200839] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1ac082cf-d4a4-4dd7-88b6-f1a56938bd69 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1322.232166] env[63088]: INFO nova.virt.vmwareapi.images [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] The imported VM was unregistered [ 1322.234599] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Caching image {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1322.234839] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Creating directory with path [datastore1] devstack-image-cache_base/5bb4b00f-be25-4207-8793-a8341056e087 {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1322.235127] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6cfed708-afb6-4b7e-a2f2-ae2f261b6dac {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1322.245416] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Created directory with path [datastore1] devstack-image-cache_base/5bb4b00f-be25-4207-8793-a8341056e087 {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1322.245608] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_cc617ad8-1565-4ed3-87d9-799f96b6381a/OSTACK_IMG_cc617ad8-1565-4ed3-87d9-799f96b6381a.vmdk to [datastore1] devstack-image-cache_base/5bb4b00f-be25-4207-8793-a8341056e087/5bb4b00f-be25-4207-8793-a8341056e087.vmdk. {{(pid=63088) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1322.245846] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-b060a317-f06d-4cf9-b527-14b30ca254d6 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1322.251968] env[63088]: DEBUG oslo_vmware.api [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1322.251968] env[63088]: value = "task-1285667" [ 1322.251968] env[63088]: _type = "Task" [ 1322.251968] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1322.260315] env[63088]: DEBUG oslo_vmware.api [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285667, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1322.762576] env[63088]: DEBUG oslo_vmware.api [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285667, 'name': MoveVirtualDisk_Task} progress is 24%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1323.264317] env[63088]: DEBUG oslo_vmware.api [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285667, 'name': MoveVirtualDisk_Task} progress is 46%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1323.764672] env[63088]: DEBUG oslo_vmware.api [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285667, 'name': MoveVirtualDisk_Task} progress is 71%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1324.264895] env[63088]: DEBUG oslo_vmware.api [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285667, 'name': MoveVirtualDisk_Task} progress is 94%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1324.764966] env[63088]: DEBUG oslo_vmware.api [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285667, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.41449} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1324.765275] env[63088]: INFO nova.virt.vmwareapi.ds_util [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_cc617ad8-1565-4ed3-87d9-799f96b6381a/OSTACK_IMG_cc617ad8-1565-4ed3-87d9-799f96b6381a.vmdk to [datastore1] devstack-image-cache_base/5bb4b00f-be25-4207-8793-a8341056e087/5bb4b00f-be25-4207-8793-a8341056e087.vmdk. [ 1324.765493] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Cleaning up location [datastore1] OSTACK_IMG_cc617ad8-1565-4ed3-87d9-799f96b6381a {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1324.765666] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_cc617ad8-1565-4ed3-87d9-799f96b6381a {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1324.765914] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f2c20cf4-9b39-4f00-a6d7-4e40d13c9f68 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1324.771925] env[63088]: DEBUG oslo_vmware.api [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1324.771925] env[63088]: value = "task-1285668" [ 1324.771925] env[63088]: _type = "Task" [ 1324.771925] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1324.779019] env[63088]: DEBUG oslo_vmware.api [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285668, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1325.282112] env[63088]: DEBUG oslo_vmware.api [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285668, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.057268} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1325.282575] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1325.282575] env[63088]: DEBUG oslo_concurrency.lockutils [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5bb4b00f-be25-4207-8793-a8341056e087/5bb4b00f-be25-4207-8793-a8341056e087.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1325.282820] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/5bb4b00f-be25-4207-8793-a8341056e087/5bb4b00f-be25-4207-8793-a8341056e087.vmdk to [datastore1] 3d23868c-b0fd-470e-970c-c9935739b7b7/3d23868c-b0fd-470e-970c-c9935739b7b7.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1325.283088] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4c870e2d-b176-46de-8da1-1f89a5ebca25 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1325.288992] env[63088]: DEBUG oslo_vmware.api [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1325.288992] env[63088]: value = "task-1285669" [ 1325.288992] env[63088]: _type = "Task" [ 1325.288992] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1325.295950] env[63088]: DEBUG oslo_vmware.api [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285669, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1325.799737] env[63088]: DEBUG oslo_vmware.api [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285669, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1326.300542] env[63088]: DEBUG oslo_vmware.api [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285669, 'name': CopyVirtualDisk_Task} progress is 46%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1326.801913] env[63088]: DEBUG oslo_vmware.api [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285669, 'name': CopyVirtualDisk_Task} progress is 69%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1327.301888] env[63088]: DEBUG oslo_vmware.api [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285669, 'name': CopyVirtualDisk_Task} progress is 91%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1327.803241] env[63088]: DEBUG oslo_vmware.api [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285669, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.18188} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1327.803449] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/5bb4b00f-be25-4207-8793-a8341056e087/5bb4b00f-be25-4207-8793-a8341056e087.vmdk to [datastore1] 3d23868c-b0fd-470e-970c-c9935739b7b7/3d23868c-b0fd-470e-970c-c9935739b7b7.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1327.804108] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0ddef02-20f7-4c40-82fa-6e81d02b5b61 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1327.825178] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Reconfiguring VM instance instance-0000006d to attach disk [datastore1] 3d23868c-b0fd-470e-970c-c9935739b7b7/3d23868c-b0fd-470e-970c-c9935739b7b7.vmdk or device None with type streamOptimized {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1327.825513] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b3ec58e5-0842-4bd0-9315-d8cbb97a303f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1327.844926] env[63088]: DEBUG oslo_vmware.api [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1327.844926] env[63088]: value = "task-1285670" [ 1327.844926] env[63088]: _type = "Task" [ 1327.844926] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1327.851075] env[63088]: DEBUG oslo_vmware.api [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285670, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1328.355563] env[63088]: DEBUG oslo_vmware.api [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285670, 'name': ReconfigVM_Task, 'duration_secs': 0.309551} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1328.355936] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Reconfigured VM instance instance-0000006d to attach disk [datastore1] 3d23868c-b0fd-470e-970c-c9935739b7b7/3d23868c-b0fd-470e-970c-c9935739b7b7.vmdk or device None with type streamOptimized {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1328.356485] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ef5c9c15-0b93-46c1-85ef-bf0645a352c1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1328.362784] env[63088]: DEBUG oslo_vmware.api [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1328.362784] env[63088]: value = "task-1285671" [ 1328.362784] env[63088]: _type = "Task" [ 1328.362784] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1328.369946] env[63088]: DEBUG oslo_vmware.api [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285671, 'name': Rename_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1328.872280] env[63088]: DEBUG oslo_vmware.api [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285671, 'name': Rename_Task, 'duration_secs': 0.141066} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1328.872598] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1328.872847] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ad33f2c5-621c-4977-81d2-f8687c07ea55 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1328.878733] env[63088]: DEBUG oslo_vmware.api [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1328.878733] env[63088]: value = "task-1285672" [ 1328.878733] env[63088]: _type = "Task" [ 1328.878733] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1328.885690] env[63088]: DEBUG oslo_vmware.api [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285672, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1329.389261] env[63088]: DEBUG oslo_vmware.api [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285672, 'name': PowerOnVM_Task, 'duration_secs': 0.452108} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1329.389652] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1329.478206] env[63088]: DEBUG nova.compute.manager [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1329.479142] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99d7e2d1-5f35-461d-990b-9ecdb98bac28 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1329.997745] env[63088]: DEBUG oslo_concurrency.lockutils [None req-138a4aa6-6db1-4e77-88bf-508d2ce7749d tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lock "3d23868c-b0fd-470e-970c-c9935739b7b7" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 18.359s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1330.853636] env[63088]: DEBUG oslo_concurrency.lockutils [None req-109f8eac-3ba7-4b5c-bd5a-7db63dc27761 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquiring lock "3d23868c-b0fd-470e-970c-c9935739b7b7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1330.853979] env[63088]: DEBUG oslo_concurrency.lockutils [None req-109f8eac-3ba7-4b5c-bd5a-7db63dc27761 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lock "3d23868c-b0fd-470e-970c-c9935739b7b7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1330.854172] env[63088]: DEBUG oslo_concurrency.lockutils [None req-109f8eac-3ba7-4b5c-bd5a-7db63dc27761 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquiring lock "3d23868c-b0fd-470e-970c-c9935739b7b7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1330.854371] env[63088]: DEBUG oslo_concurrency.lockutils [None req-109f8eac-3ba7-4b5c-bd5a-7db63dc27761 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lock "3d23868c-b0fd-470e-970c-c9935739b7b7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1330.854548] env[63088]: DEBUG oslo_concurrency.lockutils [None req-109f8eac-3ba7-4b5c-bd5a-7db63dc27761 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lock "3d23868c-b0fd-470e-970c-c9935739b7b7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1330.856732] env[63088]: INFO nova.compute.manager [None req-109f8eac-3ba7-4b5c-bd5a-7db63dc27761 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Terminating instance [ 1330.858467] env[63088]: DEBUG nova.compute.manager [None req-109f8eac-3ba7-4b5c-bd5a-7db63dc27761 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1330.858662] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-109f8eac-3ba7-4b5c-bd5a-7db63dc27761 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1330.859494] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9479352-39c9-4fb4-9497-90527957c827 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1330.867033] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-109f8eac-3ba7-4b5c-bd5a-7db63dc27761 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1330.867265] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d2006b70-16e7-410d-9a6f-b56bf0ce3a07 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1330.873599] env[63088]: DEBUG oslo_vmware.api [None req-109f8eac-3ba7-4b5c-bd5a-7db63dc27761 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1330.873599] env[63088]: value = "task-1285673" [ 1330.873599] env[63088]: _type = "Task" [ 1330.873599] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1330.880918] env[63088]: DEBUG oslo_vmware.api [None req-109f8eac-3ba7-4b5c-bd5a-7db63dc27761 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285673, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1331.383823] env[63088]: DEBUG oslo_vmware.api [None req-109f8eac-3ba7-4b5c-bd5a-7db63dc27761 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285673, 'name': PowerOffVM_Task, 'duration_secs': 0.180288} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1331.384116] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-109f8eac-3ba7-4b5c-bd5a-7db63dc27761 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1331.384287] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-109f8eac-3ba7-4b5c-bd5a-7db63dc27761 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Unregistering the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1331.384543] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bfcadd92-afa5-448d-9b92-04da053b3533 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.443993] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-109f8eac-3ba7-4b5c-bd5a-7db63dc27761 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Unregistered the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1331.444271] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-109f8eac-3ba7-4b5c-bd5a-7db63dc27761 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Deleting contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1331.444429] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-109f8eac-3ba7-4b5c-bd5a-7db63dc27761 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Deleting the datastore file [datastore1] 3d23868c-b0fd-470e-970c-c9935739b7b7 {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1331.444695] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f0a2d272-b1d9-492c-a359-db02961bb786 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.450813] env[63088]: DEBUG oslo_vmware.api [None req-109f8eac-3ba7-4b5c-bd5a-7db63dc27761 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for the task: (returnval){ [ 1331.450813] env[63088]: value = "task-1285675" [ 1331.450813] env[63088]: _type = "Task" [ 1331.450813] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1331.457937] env[63088]: DEBUG oslo_vmware.api [None req-109f8eac-3ba7-4b5c-bd5a-7db63dc27761 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285675, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1331.960303] env[63088]: DEBUG oslo_vmware.api [None req-109f8eac-3ba7-4b5c-bd5a-7db63dc27761 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Task: {'id': task-1285675, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.147404} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1331.960810] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-109f8eac-3ba7-4b5c-bd5a-7db63dc27761 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1331.960810] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-109f8eac-3ba7-4b5c-bd5a-7db63dc27761 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Deleted contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1331.960917] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-109f8eac-3ba7-4b5c-bd5a-7db63dc27761 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1331.961075] env[63088]: INFO nova.compute.manager [None req-109f8eac-3ba7-4b5c-bd5a-7db63dc27761 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1331.961324] env[63088]: DEBUG oslo.service.loopingcall [None req-109f8eac-3ba7-4b5c-bd5a-7db63dc27761 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1331.961526] env[63088]: DEBUG nova.compute.manager [-] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1331.961619] env[63088]: DEBUG nova.network.neutron [-] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1332.374607] env[63088]: DEBUG nova.compute.manager [req-ba357ca2-992c-4237-9c34-f45db2fd591d req-f505162e-d3b5-4550-9b1d-afb95e3f9d4f service nova] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Received event network-vif-deleted-e7ade0ec-6647-4782-8915-4edcb26c64e1 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1332.374667] env[63088]: INFO nova.compute.manager [req-ba357ca2-992c-4237-9c34-f45db2fd591d req-f505162e-d3b5-4550-9b1d-afb95e3f9d4f service nova] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Neutron deleted interface e7ade0ec-6647-4782-8915-4edcb26c64e1; detaching it from the instance and deleting it from the info cache [ 1332.374891] env[63088]: DEBUG nova.network.neutron [req-ba357ca2-992c-4237-9c34-f45db2fd591d req-f505162e-d3b5-4550-9b1d-afb95e3f9d4f service nova] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1332.854362] env[63088]: DEBUG nova.network.neutron [-] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1332.878244] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-47d47fbb-07de-4deb-b942-0ef6bc7ab7f1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1332.887944] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90952e4f-0c49-41ef-b0ed-b587e790aee8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1332.911064] env[63088]: DEBUG nova.compute.manager [req-ba357ca2-992c-4237-9c34-f45db2fd591d req-f505162e-d3b5-4550-9b1d-afb95e3f9d4f service nova] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Detach interface failed, port_id=e7ade0ec-6647-4782-8915-4edcb26c64e1, reason: Instance 3d23868c-b0fd-470e-970c-c9935739b7b7 could not be found. {{(pid=63088) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1333.357083] env[63088]: INFO nova.compute.manager [-] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Took 1.40 seconds to deallocate network for instance. [ 1333.863276] env[63088]: DEBUG oslo_concurrency.lockutils [None req-109f8eac-3ba7-4b5c-bd5a-7db63dc27761 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1333.863504] env[63088]: DEBUG oslo_concurrency.lockutils [None req-109f8eac-3ba7-4b5c-bd5a-7db63dc27761 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1333.863756] env[63088]: DEBUG nova.objects.instance [None req-109f8eac-3ba7-4b5c-bd5a-7db63dc27761 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lazy-loading 'resources' on Instance uuid 3d23868c-b0fd-470e-970c-c9935739b7b7 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1334.417654] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a813a6b-fbfa-4c28-a8fe-5f2414c18455 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1334.425163] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebc32b2f-22bb-4c28-8f6d-667b8cc5114c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1334.455657] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd1119f0-028e-4be0-865e-fba15a501e09 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1334.462066] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0124cc27-3d2f-418a-952d-9298e6c65762 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1334.474716] env[63088]: DEBUG nova.compute.provider_tree [None req-109f8eac-3ba7-4b5c-bd5a-7db63dc27761 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1334.978521] env[63088]: DEBUG nova.scheduler.client.report [None req-109f8eac-3ba7-4b5c-bd5a-7db63dc27761 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1335.483407] env[63088]: DEBUG oslo_concurrency.lockutils [None req-109f8eac-3ba7-4b5c-bd5a-7db63dc27761 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.620s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1335.502338] env[63088]: INFO nova.scheduler.client.report [None req-109f8eac-3ba7-4b5c-bd5a-7db63dc27761 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Deleted allocations for instance 3d23868c-b0fd-470e-970c-c9935739b7b7 [ 1335.603646] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1336.010855] env[63088]: DEBUG oslo_concurrency.lockutils [None req-109f8eac-3ba7-4b5c-bd5a-7db63dc27761 tempest-ServerActionsTestOtherB-1968194729 tempest-ServerActionsTestOtherB-1968194729-project-member] Lock "3d23868c-b0fd-470e-970c-c9935739b7b7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.157s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1338.604502] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1340.603752] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1342.605667] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1342.606103] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63088) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1343.604714] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1344.600138] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1344.603710] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1344.603871] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Starting heal instance info cache {{(pid=63088) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1345.106484] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Didn't find any instances for network info cache update. {{(pid=63088) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10012}} [ 1345.604094] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1345.604451] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager.update_available_resource {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1346.107529] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1346.107818] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1346.107918] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1346.108089] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63088) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1346.109016] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9b9d82f-57a2-4336-8bf3-d63a60ab04ad {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1346.117262] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec708167-3768-487c-ad15-f0e6d3a87890 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1346.132384] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7536837-54c8-4da0-8489-92f18d0670e1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1346.138856] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cab9e3b0-d9ca-4708-974d-9006d1f60394 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1346.166639] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180978MB free_disk=140GB free_vcpus=48 pci_devices=None {{(pid=63088) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1346.166781] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1346.166979] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1346.362350] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a2c9c100-1d85-477e-b143-9d6e275e8424 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Acquiring lock "bd55e4b9-269f-4ca4-8139-86dface352b0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1346.362639] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a2c9c100-1d85-477e-b143-9d6e275e8424 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Lock "bd55e4b9-269f-4ca4-8139-86dface352b0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1346.362851] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a2c9c100-1d85-477e-b143-9d6e275e8424 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Acquiring lock "bd55e4b9-269f-4ca4-8139-86dface352b0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1346.363050] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a2c9c100-1d85-477e-b143-9d6e275e8424 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Lock "bd55e4b9-269f-4ca4-8139-86dface352b0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1346.363233] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a2c9c100-1d85-477e-b143-9d6e275e8424 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Lock "bd55e4b9-269f-4ca4-8139-86dface352b0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1346.365435] env[63088]: INFO nova.compute.manager [None req-a2c9c100-1d85-477e-b143-9d6e275e8424 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: bd55e4b9-269f-4ca4-8139-86dface352b0] Terminating instance [ 1346.367312] env[63088]: DEBUG nova.compute.manager [None req-a2c9c100-1d85-477e-b143-9d6e275e8424 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: bd55e4b9-269f-4ca4-8139-86dface352b0] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1346.367480] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-a2c9c100-1d85-477e-b143-9d6e275e8424 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: bd55e4b9-269f-4ca4-8139-86dface352b0] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1346.368312] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59348e50-df68-4dcc-ac10-361a850f37dd {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1346.376339] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2c9c100-1d85-477e-b143-9d6e275e8424 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: bd55e4b9-269f-4ca4-8139-86dface352b0] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1346.376566] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2d16af8f-c8c6-49c3-976e-b46d951bda4a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1346.382984] env[63088]: DEBUG oslo_vmware.api [None req-a2c9c100-1d85-477e-b143-9d6e275e8424 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Waiting for the task: (returnval){ [ 1346.382984] env[63088]: value = "task-1285677" [ 1346.382984] env[63088]: _type = "Task" [ 1346.382984] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1346.390181] env[63088]: DEBUG oslo_vmware.api [None req-a2c9c100-1d85-477e-b143-9d6e275e8424 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285677, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1346.892923] env[63088]: DEBUG oslo_vmware.api [None req-a2c9c100-1d85-477e-b143-9d6e275e8424 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285677, 'name': PowerOffVM_Task, 'duration_secs': 0.192333} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1346.893283] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2c9c100-1d85-477e-b143-9d6e275e8424 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: bd55e4b9-269f-4ca4-8139-86dface352b0] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1346.893412] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-a2c9c100-1d85-477e-b143-9d6e275e8424 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: bd55e4b9-269f-4ca4-8139-86dface352b0] Unregistering the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1346.893644] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3cd4ff6e-1195-426a-b940-dfc97bdd2a9f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1346.949866] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-a2c9c100-1d85-477e-b143-9d6e275e8424 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: bd55e4b9-269f-4ca4-8139-86dface352b0] Unregistered the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1346.950132] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-a2c9c100-1d85-477e-b143-9d6e275e8424 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: bd55e4b9-269f-4ca4-8139-86dface352b0] Deleting contents of the VM from datastore datastore2 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1346.950278] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-a2c9c100-1d85-477e-b143-9d6e275e8424 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Deleting the datastore file [datastore2] bd55e4b9-269f-4ca4-8139-86dface352b0 {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1346.950543] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6506d434-f8db-494b-9b31-9f2356605a94 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1346.956273] env[63088]: DEBUG oslo_vmware.api [None req-a2c9c100-1d85-477e-b143-9d6e275e8424 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Waiting for the task: (returnval){ [ 1346.956273] env[63088]: value = "task-1285679" [ 1346.956273] env[63088]: _type = "Task" [ 1346.956273] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1346.963112] env[63088]: DEBUG oslo_vmware.api [None req-a2c9c100-1d85-477e-b143-9d6e275e8424 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285679, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1347.192415] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance 918eaae2-f821-402a-a0c2-2c140e3df4d7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1347.192641] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance bd55e4b9-269f-4ca4-8139-86dface352b0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1347.192874] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Total usable vcpus: 48, total allocated vcpus: 2 {{(pid=63088) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1347.193152] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=896MB phys_disk=200GB used_disk=2GB total_vcpus=48 used_vcpus=2 pci_stats=[] {{(pid=63088) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1347.229715] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-475b32f0-f21e-4521-a089-1c7f865cacaf {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1347.237175] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb93fbf2-21d5-4df1-9c20-4f3ecb70afa4 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1347.266805] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6aa2d81d-770b-42ed-b777-488bd22ac76f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1347.273495] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6851b1b7-6e82-41fc-a0da-b9f46417efd3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1347.286063] env[63088]: DEBUG nova.compute.provider_tree [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1347.468709] env[63088]: DEBUG oslo_vmware.api [None req-a2c9c100-1d85-477e-b143-9d6e275e8424 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285679, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.123905} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1347.469094] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-a2c9c100-1d85-477e-b143-9d6e275e8424 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1347.469357] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-a2c9c100-1d85-477e-b143-9d6e275e8424 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: bd55e4b9-269f-4ca4-8139-86dface352b0] Deleted contents of the VM from datastore datastore2 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1347.469600] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-a2c9c100-1d85-477e-b143-9d6e275e8424 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: bd55e4b9-269f-4ca4-8139-86dface352b0] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1347.469795] env[63088]: INFO nova.compute.manager [None req-a2c9c100-1d85-477e-b143-9d6e275e8424 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: bd55e4b9-269f-4ca4-8139-86dface352b0] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1347.470052] env[63088]: DEBUG oslo.service.loopingcall [None req-a2c9c100-1d85-477e-b143-9d6e275e8424 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1347.470250] env[63088]: DEBUG nova.compute.manager [-] [instance: bd55e4b9-269f-4ca4-8139-86dface352b0] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1347.470343] env[63088]: DEBUG nova.network.neutron [-] [instance: bd55e4b9-269f-4ca4-8139-86dface352b0] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1347.789747] env[63088]: DEBUG nova.scheduler.client.report [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1347.882227] env[63088]: DEBUG nova.compute.manager [req-930cf389-936f-4c92-8b3d-98d3d5ff97fc req-7266695c-cbdb-4d0a-9041-95b3b920c4cb service nova] [instance: bd55e4b9-269f-4ca4-8139-86dface352b0] Received event network-vif-deleted-b40212cc-273b-4dbb-9363-91876eec40ad {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1347.882416] env[63088]: INFO nova.compute.manager [req-930cf389-936f-4c92-8b3d-98d3d5ff97fc req-7266695c-cbdb-4d0a-9041-95b3b920c4cb service nova] [instance: bd55e4b9-269f-4ca4-8139-86dface352b0] Neutron deleted interface b40212cc-273b-4dbb-9363-91876eec40ad; detaching it from the instance and deleting it from the info cache [ 1347.882606] env[63088]: DEBUG nova.network.neutron [req-930cf389-936f-4c92-8b3d-98d3d5ff97fc req-7266695c-cbdb-4d0a-9041-95b3b920c4cb service nova] [instance: bd55e4b9-269f-4ca4-8139-86dface352b0] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1348.294977] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63088) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1348.294977] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.128s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1348.358056] env[63088]: DEBUG nova.network.neutron [-] [instance: bd55e4b9-269f-4ca4-8139-86dface352b0] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1348.385284] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-80374dae-8179-4c5c-8959-86a3a4c02e60 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1348.395418] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e66d1d7-d3c0-45ff-8436-299e6a3aa83b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1348.419402] env[63088]: DEBUG nova.compute.manager [req-930cf389-936f-4c92-8b3d-98d3d5ff97fc req-7266695c-cbdb-4d0a-9041-95b3b920c4cb service nova] [instance: bd55e4b9-269f-4ca4-8139-86dface352b0] Detach interface failed, port_id=b40212cc-273b-4dbb-9363-91876eec40ad, reason: Instance bd55e4b9-269f-4ca4-8139-86dface352b0 could not be found. {{(pid=63088) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1348.861803] env[63088]: INFO nova.compute.manager [-] [instance: bd55e4b9-269f-4ca4-8139-86dface352b0] Took 1.39 seconds to deallocate network for instance. [ 1349.368403] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a2c9c100-1d85-477e-b143-9d6e275e8424 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1349.368688] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a2c9c100-1d85-477e-b143-9d6e275e8424 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1349.368925] env[63088]: DEBUG nova.objects.instance [None req-a2c9c100-1d85-477e-b143-9d6e275e8424 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Lazy-loading 'resources' on Instance uuid bd55e4b9-269f-4ca4-8139-86dface352b0 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1349.912270] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac65a825-6e7b-4ed0-b84f-a5d2905e0159 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1349.919631] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63945424-7547-42d2-afb8-a73a6ed397b9 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1349.948864] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf51185b-380e-4b69-a4b6-3885abc60a15 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1349.955586] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60bf1907-03d0-4f50-85c5-80441e7c7f28 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1349.968007] env[63088]: DEBUG nova.compute.provider_tree [None req-a2c9c100-1d85-477e-b143-9d6e275e8424 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1350.470611] env[63088]: DEBUG nova.scheduler.client.report [None req-a2c9c100-1d85-477e-b143-9d6e275e8424 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1350.975577] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a2c9c100-1d85-477e-b143-9d6e275e8424 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.607s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1350.996747] env[63088]: INFO nova.scheduler.client.report [None req-a2c9c100-1d85-477e-b143-9d6e275e8424 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Deleted allocations for instance bd55e4b9-269f-4ca4-8139-86dface352b0 [ 1351.506056] env[63088]: DEBUG oslo_concurrency.lockutils [None req-a2c9c100-1d85-477e-b143-9d6e275e8424 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Lock "bd55e4b9-269f-4ca4-8139-86dface352b0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.143s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1352.291043] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1352.395155] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5a810e86-cc67-480d-864c-1cbf28ae1d5d tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Acquiring lock "918eaae2-f821-402a-a0c2-2c140e3df4d7" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1352.395426] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5a810e86-cc67-480d-864c-1cbf28ae1d5d tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Lock "918eaae2-f821-402a-a0c2-2c140e3df4d7" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1352.899031] env[63088]: INFO nova.compute.manager [None req-5a810e86-cc67-480d-864c-1cbf28ae1d5d tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] Detaching volume 6bdbf370-3ba7-475f-b2c7-d60435f73bd7 [ 1352.930061] env[63088]: INFO nova.virt.block_device [None req-5a810e86-cc67-480d-864c-1cbf28ae1d5d tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] Attempting to driver detach volume 6bdbf370-3ba7-475f-b2c7-d60435f73bd7 from mountpoint /dev/sdb [ 1352.930334] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a810e86-cc67-480d-864c-1cbf28ae1d5d tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] Volume detach. Driver type: vmdk {{(pid=63088) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1352.930571] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a810e86-cc67-480d-864c-1cbf28ae1d5d tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-275980', 'volume_id': '6bdbf370-3ba7-475f-b2c7-d60435f73bd7', 'name': 'volume-6bdbf370-3ba7-475f-b2c7-d60435f73bd7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '918eaae2-f821-402a-a0c2-2c140e3df4d7', 'attached_at': '', 'detached_at': '', 'volume_id': '6bdbf370-3ba7-475f-b2c7-d60435f73bd7', 'serial': '6bdbf370-3ba7-475f-b2c7-d60435f73bd7'} {{(pid=63088) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1352.931428] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3338b2df-b92a-4796-ba10-551999bb1c41 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1352.953377] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6daed81-c3b1-4a2e-9d89-02c261023c6f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1352.960483] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-614ed334-9cb4-4082-b63e-0be1fcb159b8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1352.980369] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a650ad1-7a9a-46bf-a73d-122208871160 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1352.996025] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a810e86-cc67-480d-864c-1cbf28ae1d5d tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] The volume has not been displaced from its original location: [datastore1] volume-6bdbf370-3ba7-475f-b2c7-d60435f73bd7/volume-6bdbf370-3ba7-475f-b2c7-d60435f73bd7.vmdk. No consolidation needed. {{(pid=63088) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1353.001326] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a810e86-cc67-480d-864c-1cbf28ae1d5d tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] Reconfiguring VM instance instance-0000006c to detach disk 2001 {{(pid=63088) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1353.001636] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d9916e11-8c58-41e4-8c99-69850a895271 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1353.018727] env[63088]: DEBUG oslo_vmware.api [None req-5a810e86-cc67-480d-864c-1cbf28ae1d5d tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Waiting for the task: (returnval){ [ 1353.018727] env[63088]: value = "task-1285680" [ 1353.018727] env[63088]: _type = "Task" [ 1353.018727] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1353.026263] env[63088]: DEBUG oslo_vmware.api [None req-5a810e86-cc67-480d-864c-1cbf28ae1d5d tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285680, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1353.528402] env[63088]: DEBUG oslo_vmware.api [None req-5a810e86-cc67-480d-864c-1cbf28ae1d5d tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285680, 'name': ReconfigVM_Task, 'duration_secs': 0.214211} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1353.528685] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a810e86-cc67-480d-864c-1cbf28ae1d5d tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] Reconfigured VM instance instance-0000006c to detach disk 2001 {{(pid=63088) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1353.533252] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3ba69d68-b466-4f4d-8f65-fbcb6d16ea7a {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1353.546942] env[63088]: DEBUG oslo_vmware.api [None req-5a810e86-cc67-480d-864c-1cbf28ae1d5d tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Waiting for the task: (returnval){ [ 1353.546942] env[63088]: value = "task-1285681" [ 1353.546942] env[63088]: _type = "Task" [ 1353.546942] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1353.553979] env[63088]: DEBUG oslo_vmware.api [None req-5a810e86-cc67-480d-864c-1cbf28ae1d5d tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285681, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1354.056442] env[63088]: DEBUG oslo_vmware.api [None req-5a810e86-cc67-480d-864c-1cbf28ae1d5d tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285681, 'name': ReconfigVM_Task, 'duration_secs': 0.128564} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1354.056901] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a810e86-cc67-480d-864c-1cbf28ae1d5d tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-275980', 'volume_id': '6bdbf370-3ba7-475f-b2c7-d60435f73bd7', 'name': 'volume-6bdbf370-3ba7-475f-b2c7-d60435f73bd7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '918eaae2-f821-402a-a0c2-2c140e3df4d7', 'attached_at': '', 'detached_at': '', 'volume_id': '6bdbf370-3ba7-475f-b2c7-d60435f73bd7', 'serial': '6bdbf370-3ba7-475f-b2c7-d60435f73bd7'} {{(pid=63088) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1354.596298] env[63088]: DEBUG nova.objects.instance [None req-5a810e86-cc67-480d-864c-1cbf28ae1d5d tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Lazy-loading 'flavor' on Instance uuid 918eaae2-f821-402a-a0c2-2c140e3df4d7 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1355.603453] env[63088]: DEBUG oslo_concurrency.lockutils [None req-5a810e86-cc67-480d-864c-1cbf28ae1d5d tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Lock "918eaae2-f821-402a-a0c2-2c140e3df4d7" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.208s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1356.621692] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0c37efaa-b103-4860-b973-3bfae086f8af tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Acquiring lock "918eaae2-f821-402a-a0c2-2c140e3df4d7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1356.622100] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0c37efaa-b103-4860-b973-3bfae086f8af tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Lock "918eaae2-f821-402a-a0c2-2c140e3df4d7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1356.622248] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0c37efaa-b103-4860-b973-3bfae086f8af tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Acquiring lock "918eaae2-f821-402a-a0c2-2c140e3df4d7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1356.622393] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0c37efaa-b103-4860-b973-3bfae086f8af tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Lock "918eaae2-f821-402a-a0c2-2c140e3df4d7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1356.622570] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0c37efaa-b103-4860-b973-3bfae086f8af tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Lock "918eaae2-f821-402a-a0c2-2c140e3df4d7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1356.624759] env[63088]: INFO nova.compute.manager [None req-0c37efaa-b103-4860-b973-3bfae086f8af tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] Terminating instance [ 1356.626542] env[63088]: DEBUG nova.compute.manager [None req-0c37efaa-b103-4860-b973-3bfae086f8af tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1356.626738] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-0c37efaa-b103-4860-b973-3bfae086f8af tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1356.627568] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32c89a97-2c27-4e7c-8ad7-488f50e0aeb3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1356.635035] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c37efaa-b103-4860-b973-3bfae086f8af tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1356.635261] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d1f11565-677c-459e-81c8-1a4a1fa72ed0 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1356.641733] env[63088]: DEBUG oslo_vmware.api [None req-0c37efaa-b103-4860-b973-3bfae086f8af tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Waiting for the task: (returnval){ [ 1356.641733] env[63088]: value = "task-1285682" [ 1356.641733] env[63088]: _type = "Task" [ 1356.641733] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1356.649059] env[63088]: DEBUG oslo_vmware.api [None req-0c37efaa-b103-4860-b973-3bfae086f8af tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285682, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1357.151836] env[63088]: DEBUG oslo_vmware.api [None req-0c37efaa-b103-4860-b973-3bfae086f8af tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285682, 'name': PowerOffVM_Task, 'duration_secs': 0.195312} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1357.152127] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c37efaa-b103-4860-b973-3bfae086f8af tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1357.152302] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-0c37efaa-b103-4860-b973-3bfae086f8af tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] Unregistering the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1357.152544] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6bab5a84-eca4-411d-8f66-06161f45838c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1357.209911] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-0c37efaa-b103-4860-b973-3bfae086f8af tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] Unregistered the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1357.210157] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-0c37efaa-b103-4860-b973-3bfae086f8af tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] Deleting contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1357.210346] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-0c37efaa-b103-4860-b973-3bfae086f8af tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Deleting the datastore file [datastore1] 918eaae2-f821-402a-a0c2-2c140e3df4d7 {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1357.210609] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c4909a45-3d39-48d2-b60f-338a5ae00e1c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1357.216936] env[63088]: DEBUG oslo_vmware.api [None req-0c37efaa-b103-4860-b973-3bfae086f8af tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Waiting for the task: (returnval){ [ 1357.216936] env[63088]: value = "task-1285684" [ 1357.216936] env[63088]: _type = "Task" [ 1357.216936] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1357.223857] env[63088]: DEBUG oslo_vmware.api [None req-0c37efaa-b103-4860-b973-3bfae086f8af tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285684, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1357.726480] env[63088]: DEBUG oslo_vmware.api [None req-0c37efaa-b103-4860-b973-3bfae086f8af tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285684, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.127839} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1357.726878] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-0c37efaa-b103-4860-b973-3bfae086f8af tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1357.726941] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-0c37efaa-b103-4860-b973-3bfae086f8af tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] Deleted contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1357.727110] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-0c37efaa-b103-4860-b973-3bfae086f8af tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1357.727293] env[63088]: INFO nova.compute.manager [None req-0c37efaa-b103-4860-b973-3bfae086f8af tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1357.727533] env[63088]: DEBUG oslo.service.loopingcall [None req-0c37efaa-b103-4860-b973-3bfae086f8af tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1357.727734] env[63088]: DEBUG nova.compute.manager [-] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1357.727828] env[63088]: DEBUG nova.network.neutron [-] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1358.132793] env[63088]: DEBUG nova.compute.manager [req-b4df4ca0-9aca-4911-9c4c-9104952fc31d req-0d828e4d-718e-476f-ad73-1e179c655d0b service nova] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] Received event network-vif-deleted-c1808fff-b8ad-486e-bd90-0f23d115e398 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1358.132888] env[63088]: INFO nova.compute.manager [req-b4df4ca0-9aca-4911-9c4c-9104952fc31d req-0d828e4d-718e-476f-ad73-1e179c655d0b service nova] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] Neutron deleted interface c1808fff-b8ad-486e-bd90-0f23d115e398; detaching it from the instance and deleting it from the info cache [ 1358.133057] env[63088]: DEBUG nova.network.neutron [req-b4df4ca0-9aca-4911-9c4c-9104952fc31d req-0d828e4d-718e-476f-ad73-1e179c655d0b service nova] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1358.608059] env[63088]: DEBUG nova.network.neutron [-] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1358.636626] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-134d3f39-ca56-418e-b861-44036bd93b66 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1358.646193] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87242bf0-a2b4-4159-8ff8-11c0948aeec4 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1358.668637] env[63088]: DEBUG nova.compute.manager [req-b4df4ca0-9aca-4911-9c4c-9104952fc31d req-0d828e4d-718e-476f-ad73-1e179c655d0b service nova] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] Detach interface failed, port_id=c1808fff-b8ad-486e-bd90-0f23d115e398, reason: Instance 918eaae2-f821-402a-a0c2-2c140e3df4d7 could not be found. {{(pid=63088) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1359.110573] env[63088]: INFO nova.compute.manager [-] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] Took 1.38 seconds to deallocate network for instance. [ 1359.619059] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0c37efaa-b103-4860-b973-3bfae086f8af tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1359.619059] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0c37efaa-b103-4860-b973-3bfae086f8af tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1359.619059] env[63088]: DEBUG nova.objects.instance [None req-0c37efaa-b103-4860-b973-3bfae086f8af tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Lazy-loading 'resources' on Instance uuid 918eaae2-f821-402a-a0c2-2c140e3df4d7 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1360.155716] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a63b3ae8-62af-4cfb-b0f9-185912274533 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1360.163352] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3612219-7e5d-4e54-97be-91a0a0aadcbc {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1360.192704] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6121f89c-1f2a-4056-afe6-805dc42cc9bc {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1360.199738] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-942c6028-b98d-4161-9598-03e7200309a0 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1360.213186] env[63088]: DEBUG nova.compute.provider_tree [None req-0c37efaa-b103-4860-b973-3bfae086f8af tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1360.715862] env[63088]: DEBUG nova.scheduler.client.report [None req-0c37efaa-b103-4860-b973-3bfae086f8af tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1361.221412] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0c37efaa-b103-4860-b973-3bfae086f8af tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.603s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1361.240251] env[63088]: INFO nova.scheduler.client.report [None req-0c37efaa-b103-4860-b973-3bfae086f8af tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Deleted allocations for instance 918eaae2-f821-402a-a0c2-2c140e3df4d7 [ 1361.748797] env[63088]: DEBUG oslo_concurrency.lockutils [None req-0c37efaa-b103-4860-b973-3bfae086f8af tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Lock "918eaae2-f821-402a-a0c2-2c140e3df4d7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.127s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1363.361411] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Acquiring lock "ccc0c50e-f285-4ef2-a370-5a26fc156ce2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1363.361695] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Lock "ccc0c50e-f285-4ef2-a370-5a26fc156ce2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1363.863581] env[63088]: DEBUG nova.compute.manager [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1364.382822] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1364.383124] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1364.384718] env[63088]: INFO nova.compute.claims [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1365.418157] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46e916cf-07e6-4137-a814-5e7df4ad01ae {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1365.427068] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a6c5e03-4004-4d92-8b1f-0458208fd739 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1365.455664] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2617c484-fc1a-4115-90cc-c27ace2511a9 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1365.462451] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d5e3722-3aca-4b57-b6ca-5bb64c2f09ab {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1365.474901] env[63088]: DEBUG nova.compute.provider_tree [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1365.978434] env[63088]: DEBUG nova.scheduler.client.report [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1366.483205] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.100s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1366.483754] env[63088]: DEBUG nova.compute.manager [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1366.989074] env[63088]: DEBUG nova.compute.utils [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1366.990505] env[63088]: DEBUG nova.compute.manager [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1366.990670] env[63088]: DEBUG nova.network.neutron [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1367.045806] env[63088]: DEBUG nova.policy [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '11a98d4e5ef3402b99bea85a0bb479e7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e4d8e091fab04519a6c430290830998c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 1367.309846] env[63088]: DEBUG nova.network.neutron [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] Successfully created port: 6934ea2d-1ab1-4b34-932b-d45617ed1177 {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1367.493826] env[63088]: DEBUG nova.compute.manager [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1368.505466] env[63088]: DEBUG nova.compute.manager [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1368.530260] env[63088]: DEBUG nova.virt.hardware [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1368.530524] env[63088]: DEBUG nova.virt.hardware [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1368.530688] env[63088]: DEBUG nova.virt.hardware [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1368.530874] env[63088]: DEBUG nova.virt.hardware [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1368.531034] env[63088]: DEBUG nova.virt.hardware [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1368.531195] env[63088]: DEBUG nova.virt.hardware [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1368.531403] env[63088]: DEBUG nova.virt.hardware [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1368.531562] env[63088]: DEBUG nova.virt.hardware [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1368.531731] env[63088]: DEBUG nova.virt.hardware [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1368.531911] env[63088]: DEBUG nova.virt.hardware [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1368.532122] env[63088]: DEBUG nova.virt.hardware [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1368.532978] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6143554b-cdef-4d30-b9cb-26aa67f61e3c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1368.541365] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1772cc02-afd7-464c-88ff-473d748994ff {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1368.664405] env[63088]: DEBUG nova.compute.manager [req-e953d0f1-694a-4f6f-982d-98f34cc0bc77 req-ad6e1463-ec0a-4b07-9073-525a4b65ab1d service nova] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] Received event network-vif-plugged-6934ea2d-1ab1-4b34-932b-d45617ed1177 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1368.664405] env[63088]: DEBUG oslo_concurrency.lockutils [req-e953d0f1-694a-4f6f-982d-98f34cc0bc77 req-ad6e1463-ec0a-4b07-9073-525a4b65ab1d service nova] Acquiring lock "ccc0c50e-f285-4ef2-a370-5a26fc156ce2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1368.664405] env[63088]: DEBUG oslo_concurrency.lockutils [req-e953d0f1-694a-4f6f-982d-98f34cc0bc77 req-ad6e1463-ec0a-4b07-9073-525a4b65ab1d service nova] Lock "ccc0c50e-f285-4ef2-a370-5a26fc156ce2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1368.664405] env[63088]: DEBUG oslo_concurrency.lockutils [req-e953d0f1-694a-4f6f-982d-98f34cc0bc77 req-ad6e1463-ec0a-4b07-9073-525a4b65ab1d service nova] Lock "ccc0c50e-f285-4ef2-a370-5a26fc156ce2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1368.664590] env[63088]: DEBUG nova.compute.manager [req-e953d0f1-694a-4f6f-982d-98f34cc0bc77 req-ad6e1463-ec0a-4b07-9073-525a4b65ab1d service nova] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] No waiting events found dispatching network-vif-plugged-6934ea2d-1ab1-4b34-932b-d45617ed1177 {{(pid=63088) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1368.664671] env[63088]: WARNING nova.compute.manager [req-e953d0f1-694a-4f6f-982d-98f34cc0bc77 req-ad6e1463-ec0a-4b07-9073-525a4b65ab1d service nova] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] Received unexpected event network-vif-plugged-6934ea2d-1ab1-4b34-932b-d45617ed1177 for instance with vm_state building and task_state spawning. [ 1368.743905] env[63088]: DEBUG nova.network.neutron [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] Successfully updated port: 6934ea2d-1ab1-4b34-932b-d45617ed1177 {{(pid=63088) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1369.246347] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Acquiring lock "refresh_cache-ccc0c50e-f285-4ef2-a370-5a26fc156ce2" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1369.246519] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Acquired lock "refresh_cache-ccc0c50e-f285-4ef2-a370-5a26fc156ce2" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1369.246676] env[63088]: DEBUG nova.network.neutron [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1369.777658] env[63088]: DEBUG nova.network.neutron [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1370.073094] env[63088]: DEBUG nova.network.neutron [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] Updating instance_info_cache with network_info: [{"id": "6934ea2d-1ab1-4b34-932b-d45617ed1177", "address": "fa:16:3e:c4:31:f2", "network": {"id": "ab40c372-54e5-49d6-a02a-279901071b67", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1152691158-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e4d8e091fab04519a6c430290830998c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6934ea2d-1a", "ovs_interfaceid": "6934ea2d-1ab1-4b34-932b-d45617ed1177", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1370.576300] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Releasing lock "refresh_cache-ccc0c50e-f285-4ef2-a370-5a26fc156ce2" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1370.576672] env[63088]: DEBUG nova.compute.manager [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] Instance network_info: |[{"id": "6934ea2d-1ab1-4b34-932b-d45617ed1177", "address": "fa:16:3e:c4:31:f2", "network": {"id": "ab40c372-54e5-49d6-a02a-279901071b67", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1152691158-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e4d8e091fab04519a6c430290830998c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6934ea2d-1a", "ovs_interfaceid": "6934ea2d-1ab1-4b34-932b-d45617ed1177", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1370.577219] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c4:31:f2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '418ddd3d-5f64-407e-8e0c-c8b81639bee9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6934ea2d-1ab1-4b34-932b-d45617ed1177', 'vif_model': 'vmxnet3'}] {{(pid=63088) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1370.584678] env[63088]: DEBUG oslo.service.loopingcall [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1370.584853] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] Creating VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1370.585099] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-05ced6b3-0095-4ce0-a0cc-6fc8fe0926eb {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1370.605560] env[63088]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1370.605560] env[63088]: value = "task-1285685" [ 1370.605560] env[63088]: _type = "Task" [ 1370.605560] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1370.612754] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285685, 'name': CreateVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1370.697238] env[63088]: DEBUG nova.compute.manager [req-e8e165ba-91a1-4910-b93c-9bdef28880c9 req-7b92f4d1-34b2-48df-8786-85c409fd406d service nova] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] Received event network-changed-6934ea2d-1ab1-4b34-932b-d45617ed1177 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1370.697483] env[63088]: DEBUG nova.compute.manager [req-e8e165ba-91a1-4910-b93c-9bdef28880c9 req-7b92f4d1-34b2-48df-8786-85c409fd406d service nova] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] Refreshing instance network info cache due to event network-changed-6934ea2d-1ab1-4b34-932b-d45617ed1177. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1370.697691] env[63088]: DEBUG oslo_concurrency.lockutils [req-e8e165ba-91a1-4910-b93c-9bdef28880c9 req-7b92f4d1-34b2-48df-8786-85c409fd406d service nova] Acquiring lock "refresh_cache-ccc0c50e-f285-4ef2-a370-5a26fc156ce2" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1370.697784] env[63088]: DEBUG oslo_concurrency.lockutils [req-e8e165ba-91a1-4910-b93c-9bdef28880c9 req-7b92f4d1-34b2-48df-8786-85c409fd406d service nova] Acquired lock "refresh_cache-ccc0c50e-f285-4ef2-a370-5a26fc156ce2" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1370.697953] env[63088]: DEBUG nova.network.neutron [req-e8e165ba-91a1-4910-b93c-9bdef28880c9 req-7b92f4d1-34b2-48df-8786-85c409fd406d service nova] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] Refreshing network info cache for port 6934ea2d-1ab1-4b34-932b-d45617ed1177 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1371.115151] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285685, 'name': CreateVM_Task, 'duration_secs': 0.277296} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1371.115534] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] Created VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1371.116048] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1371.116230] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1371.116556] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1371.116815] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ab22ec5e-c2ee-47d0-93f2-2f739e0d9113 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1371.121018] env[63088]: DEBUG oslo_vmware.api [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Waiting for the task: (returnval){ [ 1371.121018] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]522db398-37e3-7430-fa58-f03bd9ee4643" [ 1371.121018] env[63088]: _type = "Task" [ 1371.121018] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1371.128169] env[63088]: DEBUG oslo_vmware.api [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]522db398-37e3-7430-fa58-f03bd9ee4643, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1371.372907] env[63088]: DEBUG nova.network.neutron [req-e8e165ba-91a1-4910-b93c-9bdef28880c9 req-7b92f4d1-34b2-48df-8786-85c409fd406d service nova] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] Updated VIF entry in instance network info cache for port 6934ea2d-1ab1-4b34-932b-d45617ed1177. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1371.373311] env[63088]: DEBUG nova.network.neutron [req-e8e165ba-91a1-4910-b93c-9bdef28880c9 req-7b92f4d1-34b2-48df-8786-85c409fd406d service nova] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] Updating instance_info_cache with network_info: [{"id": "6934ea2d-1ab1-4b34-932b-d45617ed1177", "address": "fa:16:3e:c4:31:f2", "network": {"id": "ab40c372-54e5-49d6-a02a-279901071b67", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1152691158-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e4d8e091fab04519a6c430290830998c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6934ea2d-1a", "ovs_interfaceid": "6934ea2d-1ab1-4b34-932b-d45617ed1177", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1371.631641] env[63088]: DEBUG oslo_vmware.api [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]522db398-37e3-7430-fa58-f03bd9ee4643, 'name': SearchDatastore_Task, 'duration_secs': 0.010524} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1371.631894] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1371.632152] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] Processing image 1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1371.632391] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1371.632543] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1371.632724] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1371.633028] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5e71bc8c-1fd2-4761-9e97-28592e19c020 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1371.642241] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1371.642412] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1371.643089] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1316fd2e-ed02-4d86-9df8-b8130c78506e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1371.647810] env[63088]: DEBUG oslo_vmware.api [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Waiting for the task: (returnval){ [ 1371.647810] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]523532de-8c0b-0d66-d392-cc63e47c5eaa" [ 1371.647810] env[63088]: _type = "Task" [ 1371.647810] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1371.655326] env[63088]: DEBUG oslo_vmware.api [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]523532de-8c0b-0d66-d392-cc63e47c5eaa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1371.875757] env[63088]: DEBUG oslo_concurrency.lockutils [req-e8e165ba-91a1-4910-b93c-9bdef28880c9 req-7b92f4d1-34b2-48df-8786-85c409fd406d service nova] Releasing lock "refresh_cache-ccc0c50e-f285-4ef2-a370-5a26fc156ce2" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1372.158475] env[63088]: DEBUG oslo_vmware.api [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]523532de-8c0b-0d66-d392-cc63e47c5eaa, 'name': SearchDatastore_Task, 'duration_secs': 0.00781} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1372.159270] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-212d4c3e-288a-4889-bf93-5046c40498d9 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1372.164425] env[63088]: DEBUG oslo_vmware.api [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Waiting for the task: (returnval){ [ 1372.164425] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]522ad490-0566-f98b-2c8c-88a517f152a1" [ 1372.164425] env[63088]: _type = "Task" [ 1372.164425] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1372.171569] env[63088]: DEBUG oslo_vmware.api [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]522ad490-0566-f98b-2c8c-88a517f152a1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1372.675160] env[63088]: DEBUG oslo_vmware.api [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]522ad490-0566-f98b-2c8c-88a517f152a1, 'name': SearchDatastore_Task, 'duration_secs': 0.008853} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1372.675421] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1372.675683] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] ccc0c50e-f285-4ef2-a370-5a26fc156ce2/ccc0c50e-f285-4ef2-a370-5a26fc156ce2.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1372.675968] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3dcce08c-f97e-4217-b043-1515647ebe61 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1372.682288] env[63088]: DEBUG oslo_vmware.api [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Waiting for the task: (returnval){ [ 1372.682288] env[63088]: value = "task-1285686" [ 1372.682288] env[63088]: _type = "Task" [ 1372.682288] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1372.689241] env[63088]: DEBUG oslo_vmware.api [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285686, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1373.192284] env[63088]: DEBUG oslo_vmware.api [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285686, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.42095} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1373.192650] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] ccc0c50e-f285-4ef2-a370-5a26fc156ce2/ccc0c50e-f285-4ef2-a370-5a26fc156ce2.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1373.192852] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] Extending root virtual disk to 1048576 {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1373.193029] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9a8e3255-71b3-492e-9baf-8b1844ba629e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1373.199196] env[63088]: DEBUG oslo_vmware.api [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Waiting for the task: (returnval){ [ 1373.199196] env[63088]: value = "task-1285687" [ 1373.199196] env[63088]: _type = "Task" [ 1373.199196] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1373.205986] env[63088]: DEBUG oslo_vmware.api [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285687, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1373.708361] env[63088]: DEBUG oslo_vmware.api [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285687, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065295} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1373.708632] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] Extended root virtual disk {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1373.709393] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca6f426d-f9de-487d-995e-3027ae2d2eca {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1373.730252] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] Reconfiguring VM instance instance-0000006f to attach disk [datastore1] ccc0c50e-f285-4ef2-a370-5a26fc156ce2/ccc0c50e-f285-4ef2-a370-5a26fc156ce2.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1373.730480] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6e6b2edb-c356-4945-ad12-5d5905ee9337 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1373.748158] env[63088]: DEBUG oslo_vmware.api [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Waiting for the task: (returnval){ [ 1373.748158] env[63088]: value = "task-1285688" [ 1373.748158] env[63088]: _type = "Task" [ 1373.748158] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1373.755106] env[63088]: DEBUG oslo_vmware.api [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285688, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1374.259019] env[63088]: DEBUG oslo_vmware.api [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285688, 'name': ReconfigVM_Task, 'duration_secs': 0.249043} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1374.259392] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] Reconfigured VM instance instance-0000006f to attach disk [datastore1] ccc0c50e-f285-4ef2-a370-5a26fc156ce2/ccc0c50e-f285-4ef2-a370-5a26fc156ce2.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1374.259887] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1bf9aece-732e-42dd-843b-361f9856ec5c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1374.266283] env[63088]: DEBUG oslo_vmware.api [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Waiting for the task: (returnval){ [ 1374.266283] env[63088]: value = "task-1285689" [ 1374.266283] env[63088]: _type = "Task" [ 1374.266283] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1374.273582] env[63088]: DEBUG oslo_vmware.api [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285689, 'name': Rename_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1374.776445] env[63088]: DEBUG oslo_vmware.api [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285689, 'name': Rename_Task, 'duration_secs': 0.12668} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1374.776445] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1374.776445] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cb093e88-e5e7-4651-b9df-ffa34ad31367 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1374.781970] env[63088]: DEBUG oslo_vmware.api [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Waiting for the task: (returnval){ [ 1374.781970] env[63088]: value = "task-1285690" [ 1374.781970] env[63088]: _type = "Task" [ 1374.781970] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1374.788871] env[63088]: DEBUG oslo_vmware.api [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285690, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1375.292353] env[63088]: DEBUG oslo_vmware.api [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285690, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1375.792565] env[63088]: DEBUG oslo_vmware.api [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285690, 'name': PowerOnVM_Task, 'duration_secs': 0.806613} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1375.792842] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1375.793063] env[63088]: INFO nova.compute.manager [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] Took 7.29 seconds to spawn the instance on the hypervisor. [ 1375.793255] env[63088]: DEBUG nova.compute.manager [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1375.794062] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53759564-8aff-49b2-896e-87de6e1b0b95 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1376.311828] env[63088]: INFO nova.compute.manager [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] Took 11.94 seconds to build instance. [ 1376.814084] env[63088]: DEBUG oslo_concurrency.lockutils [None req-6a8f3359-9afa-439b-9953-26b0f56dddee tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Lock "ccc0c50e-f285-4ef2-a370-5a26fc156ce2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.452s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1377.443668] env[63088]: DEBUG nova.compute.manager [req-608350de-8cfe-46b4-a262-ed557de3beff req-20a0b11a-9299-4b14-978b-4d70308b2f9f service nova] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] Received event network-changed-6934ea2d-1ab1-4b34-932b-d45617ed1177 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1377.443958] env[63088]: DEBUG nova.compute.manager [req-608350de-8cfe-46b4-a262-ed557de3beff req-20a0b11a-9299-4b14-978b-4d70308b2f9f service nova] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] Refreshing instance network info cache due to event network-changed-6934ea2d-1ab1-4b34-932b-d45617ed1177. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1377.444156] env[63088]: DEBUG oslo_concurrency.lockutils [req-608350de-8cfe-46b4-a262-ed557de3beff req-20a0b11a-9299-4b14-978b-4d70308b2f9f service nova] Acquiring lock "refresh_cache-ccc0c50e-f285-4ef2-a370-5a26fc156ce2" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1377.444315] env[63088]: DEBUG oslo_concurrency.lockutils [req-608350de-8cfe-46b4-a262-ed557de3beff req-20a0b11a-9299-4b14-978b-4d70308b2f9f service nova] Acquired lock "refresh_cache-ccc0c50e-f285-4ef2-a370-5a26fc156ce2" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1377.444442] env[63088]: DEBUG nova.network.neutron [req-608350de-8cfe-46b4-a262-ed557de3beff req-20a0b11a-9299-4b14-978b-4d70308b2f9f service nova] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] Refreshing network info cache for port 6934ea2d-1ab1-4b34-932b-d45617ed1177 {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1378.149890] env[63088]: DEBUG nova.network.neutron [req-608350de-8cfe-46b4-a262-ed557de3beff req-20a0b11a-9299-4b14-978b-4d70308b2f9f service nova] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] Updated VIF entry in instance network info cache for port 6934ea2d-1ab1-4b34-932b-d45617ed1177. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1378.150294] env[63088]: DEBUG nova.network.neutron [req-608350de-8cfe-46b4-a262-ed557de3beff req-20a0b11a-9299-4b14-978b-4d70308b2f9f service nova] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] Updating instance_info_cache with network_info: [{"id": "6934ea2d-1ab1-4b34-932b-d45617ed1177", "address": "fa:16:3e:c4:31:f2", "network": {"id": "ab40c372-54e5-49d6-a02a-279901071b67", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1152691158-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.134", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e4d8e091fab04519a6c430290830998c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6934ea2d-1a", "ovs_interfaceid": "6934ea2d-1ab1-4b34-932b-d45617ed1177", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1378.653332] env[63088]: DEBUG oslo_concurrency.lockutils [req-608350de-8cfe-46b4-a262-ed557de3beff req-20a0b11a-9299-4b14-978b-4d70308b2f9f service nova] Releasing lock "refresh_cache-ccc0c50e-f285-4ef2-a370-5a26fc156ce2" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1394.604514] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1394.604948] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Cleaning up deleted instances with incomplete migration {{(pid=63088) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11264}} [ 1396.107141] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1396.604553] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1396.604738] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Cleaning up deleted instances {{(pid=63088) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11226}} [ 1397.110560] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] There are 9 instances to clean {{(pid=63088) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11235}} [ 1397.110946] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: bd55e4b9-269f-4ca4-8139-86dface352b0] Instance has had 0 of 5 cleanup attempts {{(pid=63088) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1397.614307] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: 3d23868c-b0fd-470e-970c-c9935739b7b7] Instance has had 0 of 5 cleanup attempts {{(pid=63088) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1398.118358] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: 918eaae2-f821-402a-a0c2-2c140e3df4d7] Instance has had 0 of 5 cleanup attempts {{(pid=63088) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1398.622332] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: 8e1bb264-bd9a-4779-9bbf-849c14c650a9] Instance has had 0 of 5 cleanup attempts {{(pid=63088) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1399.126082] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: f721d879-55a5-42d1-99c5-871d2e9103ec] Instance has had 0 of 5 cleanup attempts {{(pid=63088) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1399.629941] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: 7d8e3c6f-0328-405e-a869-d014caad8641] Instance has had 0 of 5 cleanup attempts {{(pid=63088) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1400.133207] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: 3fa11db4-2f70-495e-a562-ac5f06153847] Instance has had 0 of 5 cleanup attempts {{(pid=63088) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1400.637371] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: b25b5883-20c0-4cc9-8cc1-2c22a0f5be81] Instance has had 0 of 5 cleanup attempts {{(pid=63088) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1401.141609] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: 2b87a9f1-5f10-43c2-8bc9-6d560ec88015] Instance has had 0 of 5 cleanup attempts {{(pid=63088) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1404.644597] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1404.644998] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1404.644998] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Starting heal instance info cache {{(pid=63088) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1404.645104] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Rebuilding the list of instances to heal {{(pid=63088) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 1405.174813] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Acquiring lock "refresh_cache-ccc0c50e-f285-4ef2-a370-5a26fc156ce2" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1405.174953] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Acquired lock "refresh_cache-ccc0c50e-f285-4ef2-a370-5a26fc156ce2" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1405.175114] env[63088]: DEBUG nova.network.neutron [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] Forcefully refreshing network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1405.175279] env[63088]: DEBUG nova.objects.instance [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Lazy-loading 'info_cache' on Instance uuid ccc0c50e-f285-4ef2-a370-5a26fc156ce2 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1406.886671] env[63088]: DEBUG nova.network.neutron [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] Updating instance_info_cache with network_info: [{"id": "6934ea2d-1ab1-4b34-932b-d45617ed1177", "address": "fa:16:3e:c4:31:f2", "network": {"id": "ab40c372-54e5-49d6-a02a-279901071b67", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1152691158-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.134", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e4d8e091fab04519a6c430290830998c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6934ea2d-1a", "ovs_interfaceid": "6934ea2d-1ab1-4b34-932b-d45617ed1177", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1407.389687] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Releasing lock "refresh_cache-ccc0c50e-f285-4ef2-a370-5a26fc156ce2" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1407.389939] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] Updated the network info_cache for instance {{(pid=63088) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 1407.390138] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1407.390313] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1407.390461] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1407.390609] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1407.390759] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1407.390886] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63088) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1407.604244] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager.update_available_resource {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1408.107541] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1408.108049] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1408.108049] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1408.108169] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63088) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1408.109423] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4fc50d1-76b7-4f24-a9b2-4a4731a38b13 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1408.117574] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c06ffd9e-b153-4113-9932-726764f5ab9c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1408.131237] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47692083-f336-4da4-8d6b-283fc712b8dc {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1408.137339] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-123550e7-8fc3-4937-a6cf-d645a97716b4 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1408.164820] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181293MB free_disk=140GB free_vcpus=48 pci_devices=None {{(pid=63088) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1408.164984] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1408.165172] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1409.277267] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance ccc0c50e-f285-4ef2-a370-5a26fc156ce2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1409.277563] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=63088) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1409.277629] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=63088) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1409.302327] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e47510ca-bd7b-4388-a113-2af545646e17 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1409.309821] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f00444c-e8f0-405c-b379-4f8130dc26a7 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1409.338891] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56a230b8-a4b6-4607-a05a-172c2ca706ba {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1409.345247] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74f0acd4-5953-44ea-9d72-7b044d97de0e {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1409.357451] env[63088]: DEBUG nova.compute.provider_tree [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1409.860730] env[63088]: DEBUG nova.scheduler.client.report [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1410.366619] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63088) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1410.367068] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.202s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1412.604873] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1414.219978] env[63088]: DEBUG oslo_concurrency.lockutils [None req-29b0f291-eb39-44c8-9791-43902c9233aa tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Acquiring lock "ccc0c50e-f285-4ef2-a370-5a26fc156ce2" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1414.220268] env[63088]: DEBUG oslo_concurrency.lockutils [None req-29b0f291-eb39-44c8-9791-43902c9233aa tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Lock "ccc0c50e-f285-4ef2-a370-5a26fc156ce2" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1414.723446] env[63088]: DEBUG nova.compute.utils [None req-29b0f291-eb39-44c8-9791-43902c9233aa tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1415.226187] env[63088]: DEBUG oslo_concurrency.lockutils [None req-29b0f291-eb39-44c8-9791-43902c9233aa tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Lock "ccc0c50e-f285-4ef2-a370-5a26fc156ce2" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1416.286051] env[63088]: DEBUG oslo_concurrency.lockutils [None req-29b0f291-eb39-44c8-9791-43902c9233aa tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Acquiring lock "ccc0c50e-f285-4ef2-a370-5a26fc156ce2" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1416.286051] env[63088]: DEBUG oslo_concurrency.lockutils [None req-29b0f291-eb39-44c8-9791-43902c9233aa tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Lock "ccc0c50e-f285-4ef2-a370-5a26fc156ce2" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1416.286051] env[63088]: INFO nova.compute.manager [None req-29b0f291-eb39-44c8-9791-43902c9233aa tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] Attaching volume 56a48c02-0100-4c67-a369-21cc28c6a3ef to /dev/sdb [ 1416.315144] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23c7b24d-baae-4f60-aa3e-d8d77a818cae {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1416.322104] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba23b0fe-10bd-42ab-8d6a-ff0efc224647 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1416.334336] env[63088]: DEBUG nova.virt.block_device [None req-29b0f291-eb39-44c8-9791-43902c9233aa tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] Updating existing volume attachment record: 0bbe266b-6bf2-4970-8a6d-f9dbace363b6 {{(pid=63088) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1420.875016] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-29b0f291-eb39-44c8-9791-43902c9233aa tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] Volume attach. Driver type: vmdk {{(pid=63088) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1420.875287] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-29b0f291-eb39-44c8-9791-43902c9233aa tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-275987', 'volume_id': '56a48c02-0100-4c67-a369-21cc28c6a3ef', 'name': 'volume-56a48c02-0100-4c67-a369-21cc28c6a3ef', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ccc0c50e-f285-4ef2-a370-5a26fc156ce2', 'attached_at': '', 'detached_at': '', 'volume_id': '56a48c02-0100-4c67-a369-21cc28c6a3ef', 'serial': '56a48c02-0100-4c67-a369-21cc28c6a3ef'} {{(pid=63088) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1420.876188] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3629c05-bd21-46a5-8856-bf70b19a88ec {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1420.891918] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd6437ba-e52f-4db4-b0b3-8b81b2b243f7 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1420.915179] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-29b0f291-eb39-44c8-9791-43902c9233aa tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] Reconfiguring VM instance instance-0000006f to attach disk [datastore1] volume-56a48c02-0100-4c67-a369-21cc28c6a3ef/volume-56a48c02-0100-4c67-a369-21cc28c6a3ef.vmdk or device None with type thin {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1420.915421] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6090b435-649e-40fa-abf9-5f606ef88853 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1420.931964] env[63088]: DEBUG oslo_vmware.api [None req-29b0f291-eb39-44c8-9791-43902c9233aa tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Waiting for the task: (returnval){ [ 1420.931964] env[63088]: value = "task-1285693" [ 1420.931964] env[63088]: _type = "Task" [ 1420.931964] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1420.939688] env[63088]: DEBUG oslo_vmware.api [None req-29b0f291-eb39-44c8-9791-43902c9233aa tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285693, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1421.441356] env[63088]: DEBUG oslo_vmware.api [None req-29b0f291-eb39-44c8-9791-43902c9233aa tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285693, 'name': ReconfigVM_Task, 'duration_secs': 0.332294} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1421.441642] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-29b0f291-eb39-44c8-9791-43902c9233aa tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] Reconfigured VM instance instance-0000006f to attach disk [datastore1] volume-56a48c02-0100-4c67-a369-21cc28c6a3ef/volume-56a48c02-0100-4c67-a369-21cc28c6a3ef.vmdk or device None with type thin {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1421.446288] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-905a8ab5-b1bc-4a24-82cd-c952102543d2 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1421.460617] env[63088]: DEBUG oslo_vmware.api [None req-29b0f291-eb39-44c8-9791-43902c9233aa tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Waiting for the task: (returnval){ [ 1421.460617] env[63088]: value = "task-1285694" [ 1421.460617] env[63088]: _type = "Task" [ 1421.460617] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1421.467900] env[63088]: DEBUG oslo_vmware.api [None req-29b0f291-eb39-44c8-9791-43902c9233aa tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285694, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1421.970717] env[63088]: DEBUG oslo_vmware.api [None req-29b0f291-eb39-44c8-9791-43902c9233aa tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285694, 'name': ReconfigVM_Task, 'duration_secs': 0.12748} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1421.971029] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-29b0f291-eb39-44c8-9791-43902c9233aa tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-275987', 'volume_id': '56a48c02-0100-4c67-a369-21cc28c6a3ef', 'name': 'volume-56a48c02-0100-4c67-a369-21cc28c6a3ef', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ccc0c50e-f285-4ef2-a370-5a26fc156ce2', 'attached_at': '', 'detached_at': '', 'volume_id': '56a48c02-0100-4c67-a369-21cc28c6a3ef', 'serial': '56a48c02-0100-4c67-a369-21cc28c6a3ef'} {{(pid=63088) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1423.006880] env[63088]: DEBUG nova.objects.instance [None req-29b0f291-eb39-44c8-9791-43902c9233aa tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Lazy-loading 'flavor' on Instance uuid ccc0c50e-f285-4ef2-a370-5a26fc156ce2 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1423.513789] env[63088]: DEBUG oslo_concurrency.lockutils [None req-29b0f291-eb39-44c8-9791-43902c9233aa tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Lock "ccc0c50e-f285-4ef2-a370-5a26fc156ce2" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.228s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1423.717171] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f2dc96f1-4c3b-4ad0-a92c-97551b90069b tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Acquiring lock "ccc0c50e-f285-4ef2-a370-5a26fc156ce2" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1423.717542] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f2dc96f1-4c3b-4ad0-a92c-97551b90069b tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Lock "ccc0c50e-f285-4ef2-a370-5a26fc156ce2" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1424.220572] env[63088]: INFO nova.compute.manager [None req-f2dc96f1-4c3b-4ad0-a92c-97551b90069b tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] Detaching volume 56a48c02-0100-4c67-a369-21cc28c6a3ef [ 1424.250347] env[63088]: INFO nova.virt.block_device [None req-f2dc96f1-4c3b-4ad0-a92c-97551b90069b tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] Attempting to driver detach volume 56a48c02-0100-4c67-a369-21cc28c6a3ef from mountpoint /dev/sdb [ 1424.250589] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-f2dc96f1-4c3b-4ad0-a92c-97551b90069b tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] Volume detach. Driver type: vmdk {{(pid=63088) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1424.250778] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-f2dc96f1-4c3b-4ad0-a92c-97551b90069b tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-275987', 'volume_id': '56a48c02-0100-4c67-a369-21cc28c6a3ef', 'name': 'volume-56a48c02-0100-4c67-a369-21cc28c6a3ef', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ccc0c50e-f285-4ef2-a370-5a26fc156ce2', 'attached_at': '', 'detached_at': '', 'volume_id': '56a48c02-0100-4c67-a369-21cc28c6a3ef', 'serial': '56a48c02-0100-4c67-a369-21cc28c6a3ef'} {{(pid=63088) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1424.251655] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb418f92-d162-4a35-bc07-e017d5513c34 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1424.272372] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96c27994-256c-4e28-bdcb-872a1861f854 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1424.278697] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5fe5f8f-3721-482c-a046-f0a3aae68f8b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1424.297873] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e6d7064-7e28-4720-b155-fb4d502288b8 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1424.311696] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-f2dc96f1-4c3b-4ad0-a92c-97551b90069b tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] The volume has not been displaced from its original location: [datastore1] volume-56a48c02-0100-4c67-a369-21cc28c6a3ef/volume-56a48c02-0100-4c67-a369-21cc28c6a3ef.vmdk. No consolidation needed. {{(pid=63088) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1424.316949] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-f2dc96f1-4c3b-4ad0-a92c-97551b90069b tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] Reconfiguring VM instance instance-0000006f to detach disk 2001 {{(pid=63088) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1424.317217] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9047d151-9921-4c04-a593-31861722e3ec {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1424.333445] env[63088]: DEBUG oslo_vmware.api [None req-f2dc96f1-4c3b-4ad0-a92c-97551b90069b tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Waiting for the task: (returnval){ [ 1424.333445] env[63088]: value = "task-1285695" [ 1424.333445] env[63088]: _type = "Task" [ 1424.333445] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1424.341618] env[63088]: DEBUG oslo_vmware.api [None req-f2dc96f1-4c3b-4ad0-a92c-97551b90069b tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285695, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1424.843053] env[63088]: DEBUG oslo_vmware.api [None req-f2dc96f1-4c3b-4ad0-a92c-97551b90069b tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285695, 'name': ReconfigVM_Task, 'duration_secs': 0.203484} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1424.843329] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-f2dc96f1-4c3b-4ad0-a92c-97551b90069b tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] Reconfigured VM instance instance-0000006f to detach disk 2001 {{(pid=63088) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1424.848008] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-93e23eb9-5905-4996-a2b4-6e86a9f4900f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1424.862296] env[63088]: DEBUG oslo_vmware.api [None req-f2dc96f1-4c3b-4ad0-a92c-97551b90069b tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Waiting for the task: (returnval){ [ 1424.862296] env[63088]: value = "task-1285696" [ 1424.862296] env[63088]: _type = "Task" [ 1424.862296] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1424.869479] env[63088]: DEBUG oslo_vmware.api [None req-f2dc96f1-4c3b-4ad0-a92c-97551b90069b tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285696, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1425.371708] env[63088]: DEBUG oslo_vmware.api [None req-f2dc96f1-4c3b-4ad0-a92c-97551b90069b tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285696, 'name': ReconfigVM_Task, 'duration_secs': 0.135052} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1425.372052] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-f2dc96f1-4c3b-4ad0-a92c-97551b90069b tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-275987', 'volume_id': '56a48c02-0100-4c67-a369-21cc28c6a3ef', 'name': 'volume-56a48c02-0100-4c67-a369-21cc28c6a3ef', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ccc0c50e-f285-4ef2-a370-5a26fc156ce2', 'attached_at': '', 'detached_at': '', 'volume_id': '56a48c02-0100-4c67-a369-21cc28c6a3ef', 'serial': '56a48c02-0100-4c67-a369-21cc28c6a3ef'} {{(pid=63088) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1425.911251] env[63088]: DEBUG nova.objects.instance [None req-f2dc96f1-4c3b-4ad0-a92c-97551b90069b tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Lazy-loading 'flavor' on Instance uuid ccc0c50e-f285-4ef2-a370-5a26fc156ce2 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1426.918330] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f2dc96f1-4c3b-4ad0-a92c-97551b90069b tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Lock "ccc0c50e-f285-4ef2-a370-5a26fc156ce2" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.201s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1427.946701] env[63088]: DEBUG oslo_concurrency.lockutils [None req-43f891f9-1eb4-430b-af60-149974e17603 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Acquiring lock "ccc0c50e-f285-4ef2-a370-5a26fc156ce2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1427.947165] env[63088]: DEBUG oslo_concurrency.lockutils [None req-43f891f9-1eb4-430b-af60-149974e17603 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Lock "ccc0c50e-f285-4ef2-a370-5a26fc156ce2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1427.947248] env[63088]: DEBUG oslo_concurrency.lockutils [None req-43f891f9-1eb4-430b-af60-149974e17603 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Acquiring lock "ccc0c50e-f285-4ef2-a370-5a26fc156ce2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1427.947442] env[63088]: DEBUG oslo_concurrency.lockutils [None req-43f891f9-1eb4-430b-af60-149974e17603 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Lock "ccc0c50e-f285-4ef2-a370-5a26fc156ce2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1427.947653] env[63088]: DEBUG oslo_concurrency.lockutils [None req-43f891f9-1eb4-430b-af60-149974e17603 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Lock "ccc0c50e-f285-4ef2-a370-5a26fc156ce2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1427.949774] env[63088]: INFO nova.compute.manager [None req-43f891f9-1eb4-430b-af60-149974e17603 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] Terminating instance [ 1427.951499] env[63088]: DEBUG nova.compute.manager [None req-43f891f9-1eb4-430b-af60-149974e17603 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1427.951699] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-43f891f9-1eb4-430b-af60-149974e17603 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1427.952525] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cdea901-7fbf-46cc-bf5e-c2d0ca9499fc {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1427.960354] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-43f891f9-1eb4-430b-af60-149974e17603 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1427.960832] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b3f63114-540c-4686-aa2e-b9b7a2620ebe {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1427.967090] env[63088]: DEBUG oslo_vmware.api [None req-43f891f9-1eb4-430b-af60-149974e17603 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Waiting for the task: (returnval){ [ 1427.967090] env[63088]: value = "task-1285697" [ 1427.967090] env[63088]: _type = "Task" [ 1427.967090] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1427.974850] env[63088]: DEBUG oslo_vmware.api [None req-43f891f9-1eb4-430b-af60-149974e17603 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285697, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1428.477254] env[63088]: DEBUG oslo_vmware.api [None req-43f891f9-1eb4-430b-af60-149974e17603 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285697, 'name': PowerOffVM_Task, 'duration_secs': 0.168104} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1428.477543] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-43f891f9-1eb4-430b-af60-149974e17603 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1428.477717] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-43f891f9-1eb4-430b-af60-149974e17603 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] Unregistering the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1428.477962] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bc47a406-971f-4b19-8025-27c49e019417 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1428.536670] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-43f891f9-1eb4-430b-af60-149974e17603 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] Unregistered the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1428.536940] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-43f891f9-1eb4-430b-af60-149974e17603 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] Deleting contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1428.537103] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-43f891f9-1eb4-430b-af60-149974e17603 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Deleting the datastore file [datastore1] ccc0c50e-f285-4ef2-a370-5a26fc156ce2 {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1428.537433] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-53b34aa1-efee-4c79-a29c-a4ce2d2af581 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1428.543912] env[63088]: DEBUG oslo_vmware.api [None req-43f891f9-1eb4-430b-af60-149974e17603 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Waiting for the task: (returnval){ [ 1428.543912] env[63088]: value = "task-1285699" [ 1428.543912] env[63088]: _type = "Task" [ 1428.543912] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1428.551105] env[63088]: DEBUG oslo_vmware.api [None req-43f891f9-1eb4-430b-af60-149974e17603 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285699, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1429.053942] env[63088]: DEBUG oslo_vmware.api [None req-43f891f9-1eb4-430b-af60-149974e17603 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285699, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.140126} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1429.054314] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-43f891f9-1eb4-430b-af60-149974e17603 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1429.054368] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-43f891f9-1eb4-430b-af60-149974e17603 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] Deleted contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1429.054572] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-43f891f9-1eb4-430b-af60-149974e17603 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1429.054698] env[63088]: INFO nova.compute.manager [None req-43f891f9-1eb4-430b-af60-149974e17603 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1429.054934] env[63088]: DEBUG oslo.service.loopingcall [None req-43f891f9-1eb4-430b-af60-149974e17603 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1429.055167] env[63088]: DEBUG nova.compute.manager [-] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1429.055277] env[63088]: DEBUG nova.network.neutron [-] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1429.464204] env[63088]: DEBUG nova.compute.manager [req-7048574c-bb88-4634-8bac-2a3c446a5fb6 req-8f7c484c-f740-4965-86bd-b4383c3dfa17 service nova] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] Received event network-vif-deleted-6934ea2d-1ab1-4b34-932b-d45617ed1177 {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1429.464425] env[63088]: INFO nova.compute.manager [req-7048574c-bb88-4634-8bac-2a3c446a5fb6 req-8f7c484c-f740-4965-86bd-b4383c3dfa17 service nova] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] Neutron deleted interface 6934ea2d-1ab1-4b34-932b-d45617ed1177; detaching it from the instance and deleting it from the info cache [ 1429.464619] env[63088]: DEBUG nova.network.neutron [req-7048574c-bb88-4634-8bac-2a3c446a5fb6 req-8f7c484c-f740-4965-86bd-b4383c3dfa17 service nova] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1429.939839] env[63088]: DEBUG nova.network.neutron [-] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1429.967762] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c2f829ae-fe8f-433b-b419-6f5b7853acfa {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1429.976866] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55cb2ad8-2843-4082-ba89-7699992163f9 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1430.002550] env[63088]: DEBUG nova.compute.manager [req-7048574c-bb88-4634-8bac-2a3c446a5fb6 req-8f7c484c-f740-4965-86bd-b4383c3dfa17 service nova] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] Detach interface failed, port_id=6934ea2d-1ab1-4b34-932b-d45617ed1177, reason: Instance ccc0c50e-f285-4ef2-a370-5a26fc156ce2 could not be found. {{(pid=63088) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1430.442439] env[63088]: INFO nova.compute.manager [-] [instance: ccc0c50e-f285-4ef2-a370-5a26fc156ce2] Took 1.39 seconds to deallocate network for instance. [ 1430.949259] env[63088]: DEBUG oslo_concurrency.lockutils [None req-43f891f9-1eb4-430b-af60-149974e17603 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1430.949541] env[63088]: DEBUG oslo_concurrency.lockutils [None req-43f891f9-1eb4-430b-af60-149974e17603 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1430.949774] env[63088]: DEBUG nova.objects.instance [None req-43f891f9-1eb4-430b-af60-149974e17603 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Lazy-loading 'resources' on Instance uuid ccc0c50e-f285-4ef2-a370-5a26fc156ce2 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1431.484398] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-361d655a-f371-428f-a318-6261d9af3ff4 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1431.491723] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3b22e92-251a-4edb-bdc0-caac2d00ac3b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1431.521734] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3b2a0d3-d79c-4abe-8d62-3600de110fa7 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1431.528927] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f82092fb-7c2e-45f9-8b2a-b516e254ea6b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1431.542852] env[63088]: DEBUG nova.compute.provider_tree [None req-43f891f9-1eb4-430b-af60-149974e17603 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1432.045802] env[63088]: DEBUG nova.scheduler.client.report [None req-43f891f9-1eb4-430b-af60-149974e17603 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1432.550970] env[63088]: DEBUG oslo_concurrency.lockutils [None req-43f891f9-1eb4-430b-af60-149974e17603 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.601s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1432.572719] env[63088]: INFO nova.scheduler.client.report [None req-43f891f9-1eb4-430b-af60-149974e17603 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Deleted allocations for instance ccc0c50e-f285-4ef2-a370-5a26fc156ce2 [ 1433.080960] env[63088]: DEBUG oslo_concurrency.lockutils [None req-43f891f9-1eb4-430b-af60-149974e17603 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Lock "ccc0c50e-f285-4ef2-a370-5a26fc156ce2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.134s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1434.665812] env[63088]: DEBUG oslo_concurrency.lockutils [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Acquiring lock "7af8d406-b000-4e0c-a230-69d114672eb5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1434.666132] env[63088]: DEBUG oslo_concurrency.lockutils [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Lock "7af8d406-b000-4e0c-a230-69d114672eb5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1435.168099] env[63088]: DEBUG nova.compute.manager [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] Starting instance... {{(pid=63088) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1435.687829] env[63088]: DEBUG oslo_concurrency.lockutils [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1435.688152] env[63088]: DEBUG oslo_concurrency.lockutils [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1435.689723] env[63088]: INFO nova.compute.claims [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1436.724711] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b752fcc-a7fe-4e0c-8e62-67769c8b0602 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1436.732419] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d75a091-b574-4692-abfd-a2825b10a8d0 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1436.762244] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3622ae0-5672-4e8c-8729-ab98c6b7878d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1436.769380] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e90af582-931b-40d5-b1ba-b68a90a3f126 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1436.782113] env[63088]: DEBUG nova.compute.provider_tree [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1437.285313] env[63088]: DEBUG nova.scheduler.client.report [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1437.790058] env[63088]: DEBUG oslo_concurrency.lockutils [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.102s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1437.790465] env[63088]: DEBUG nova.compute.manager [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] Start building networks asynchronously for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1438.295261] env[63088]: DEBUG nova.compute.utils [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1438.296752] env[63088]: DEBUG nova.compute.manager [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] Allocating IP information in the background. {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1438.296926] env[63088]: DEBUG nova.network.neutron [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] allocate_for_instance() {{(pid=63088) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1438.351699] env[63088]: DEBUG nova.policy [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '11a98d4e5ef3402b99bea85a0bb479e7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e4d8e091fab04519a6c430290830998c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63088) authorize /opt/stack/nova/nova/policy.py:201}} [ 1438.603312] env[63088]: DEBUG nova.network.neutron [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] Successfully created port: ab8e9f5f-b27c-4b87-a4a4-80c296412a6b {{(pid=63088) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1438.800887] env[63088]: DEBUG nova.compute.manager [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] Start building block device mappings for instance. {{(pid=63088) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1439.811261] env[63088]: DEBUG nova.compute.manager [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] Start spawning the instance on the hypervisor. {{(pid=63088) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1439.836367] env[63088]: DEBUG nova.virt.hardware [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T12:12:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T12:12:13Z,direct_url=,disk_format='vmdk',id=1e8c5d18-0a03-4e18-afe1-de5a6e255953,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bfb6bc5062fc4d51af1d3c577659be9a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T12:12:14Z,virtual_size=,visibility=), allow threads: False {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1439.836658] env[63088]: DEBUG nova.virt.hardware [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Flavor limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1439.836846] env[63088]: DEBUG nova.virt.hardware [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Image limits 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1439.837211] env[63088]: DEBUG nova.virt.hardware [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Flavor pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1439.837211] env[63088]: DEBUG nova.virt.hardware [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Image pref 0:0:0 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1439.837364] env[63088]: DEBUG nova.virt.hardware [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63088) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1439.837577] env[63088]: DEBUG nova.virt.hardware [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1439.837752] env[63088]: DEBUG nova.virt.hardware [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1439.838057] env[63088]: DEBUG nova.virt.hardware [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Got 1 possible topologies {{(pid=63088) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1439.838131] env[63088]: DEBUG nova.virt.hardware [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1439.838272] env[63088]: DEBUG nova.virt.hardware [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63088) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1439.839145] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98c03efc-05b9-446e-a2a0-3c115d553711 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1439.846778] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad72ac4d-e1ef-4aad-8cba-cf4e71a422ea {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1439.960690] env[63088]: DEBUG nova.compute.manager [req-af9bfa6a-0d12-40fa-b830-2498fff5bb01 req-06d18dd6-cbd8-4c9b-840f-4c2893be9b19 service nova] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] Received event network-vif-plugged-ab8e9f5f-b27c-4b87-a4a4-80c296412a6b {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1439.960921] env[63088]: DEBUG oslo_concurrency.lockutils [req-af9bfa6a-0d12-40fa-b830-2498fff5bb01 req-06d18dd6-cbd8-4c9b-840f-4c2893be9b19 service nova] Acquiring lock "7af8d406-b000-4e0c-a230-69d114672eb5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1439.961145] env[63088]: DEBUG oslo_concurrency.lockutils [req-af9bfa6a-0d12-40fa-b830-2498fff5bb01 req-06d18dd6-cbd8-4c9b-840f-4c2893be9b19 service nova] Lock "7af8d406-b000-4e0c-a230-69d114672eb5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1439.961315] env[63088]: DEBUG oslo_concurrency.lockutils [req-af9bfa6a-0d12-40fa-b830-2498fff5bb01 req-06d18dd6-cbd8-4c9b-840f-4c2893be9b19 service nova] Lock "7af8d406-b000-4e0c-a230-69d114672eb5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1439.961486] env[63088]: DEBUG nova.compute.manager [req-af9bfa6a-0d12-40fa-b830-2498fff5bb01 req-06d18dd6-cbd8-4c9b-840f-4c2893be9b19 service nova] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] No waiting events found dispatching network-vif-plugged-ab8e9f5f-b27c-4b87-a4a4-80c296412a6b {{(pid=63088) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1439.961657] env[63088]: WARNING nova.compute.manager [req-af9bfa6a-0d12-40fa-b830-2498fff5bb01 req-06d18dd6-cbd8-4c9b-840f-4c2893be9b19 service nova] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] Received unexpected event network-vif-plugged-ab8e9f5f-b27c-4b87-a4a4-80c296412a6b for instance with vm_state building and task_state spawning. [ 1440.042706] env[63088]: DEBUG nova.network.neutron [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] Successfully updated port: ab8e9f5f-b27c-4b87-a4a4-80c296412a6b {{(pid=63088) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1440.547745] env[63088]: DEBUG oslo_concurrency.lockutils [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Acquiring lock "refresh_cache-7af8d406-b000-4e0c-a230-69d114672eb5" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1440.547849] env[63088]: DEBUG oslo_concurrency.lockutils [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Acquired lock "refresh_cache-7af8d406-b000-4e0c-a230-69d114672eb5" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1440.548024] env[63088]: DEBUG nova.network.neutron [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] Building network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1441.081267] env[63088]: DEBUG nova.network.neutron [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] Instance cache missing network info. {{(pid=63088) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1441.197864] env[63088]: DEBUG nova.network.neutron [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] Updating instance_info_cache with network_info: [{"id": "ab8e9f5f-b27c-4b87-a4a4-80c296412a6b", "address": "fa:16:3e:7f:9c:6b", "network": {"id": "ab40c372-54e5-49d6-a02a-279901071b67", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1152691158-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e4d8e091fab04519a6c430290830998c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapab8e9f5f-b2", "ovs_interfaceid": "ab8e9f5f-b27c-4b87-a4a4-80c296412a6b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1441.701156] env[63088]: DEBUG oslo_concurrency.lockutils [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Releasing lock "refresh_cache-7af8d406-b000-4e0c-a230-69d114672eb5" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1441.701514] env[63088]: DEBUG nova.compute.manager [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] Instance network_info: |[{"id": "ab8e9f5f-b27c-4b87-a4a4-80c296412a6b", "address": "fa:16:3e:7f:9c:6b", "network": {"id": "ab40c372-54e5-49d6-a02a-279901071b67", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1152691158-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e4d8e091fab04519a6c430290830998c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapab8e9f5f-b2", "ovs_interfaceid": "ab8e9f5f-b27c-4b87-a4a4-80c296412a6b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63088) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1441.701976] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7f:9c:6b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '418ddd3d-5f64-407e-8e0c-c8b81639bee9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ab8e9f5f-b27c-4b87-a4a4-80c296412a6b', 'vif_model': 'vmxnet3'}] {{(pid=63088) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1441.710614] env[63088]: DEBUG oslo.service.loopingcall [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1441.710844] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] Creating VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1441.711082] env[63088]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-87e0c6c0-83f3-408b-8d4d-9ff24dc642ca {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1441.732920] env[63088]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1441.732920] env[63088]: value = "task-1285700" [ 1441.732920] env[63088]: _type = "Task" [ 1441.732920] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1441.740450] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285700, 'name': CreateVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1441.997090] env[63088]: DEBUG nova.compute.manager [req-511e4ae3-abc9-4506-8da8-17ff98f5d0ab req-738ec0ec-084e-4baf-a427-ef6075ab850d service nova] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] Received event network-changed-ab8e9f5f-b27c-4b87-a4a4-80c296412a6b {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1441.997380] env[63088]: DEBUG nova.compute.manager [req-511e4ae3-abc9-4506-8da8-17ff98f5d0ab req-738ec0ec-084e-4baf-a427-ef6075ab850d service nova] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] Refreshing instance network info cache due to event network-changed-ab8e9f5f-b27c-4b87-a4a4-80c296412a6b. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1441.997380] env[63088]: DEBUG oslo_concurrency.lockutils [req-511e4ae3-abc9-4506-8da8-17ff98f5d0ab req-738ec0ec-084e-4baf-a427-ef6075ab850d service nova] Acquiring lock "refresh_cache-7af8d406-b000-4e0c-a230-69d114672eb5" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1441.997558] env[63088]: DEBUG oslo_concurrency.lockutils [req-511e4ae3-abc9-4506-8da8-17ff98f5d0ab req-738ec0ec-084e-4baf-a427-ef6075ab850d service nova] Acquired lock "refresh_cache-7af8d406-b000-4e0c-a230-69d114672eb5" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1441.997675] env[63088]: DEBUG nova.network.neutron [req-511e4ae3-abc9-4506-8da8-17ff98f5d0ab req-738ec0ec-084e-4baf-a427-ef6075ab850d service nova] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] Refreshing network info cache for port ab8e9f5f-b27c-4b87-a4a4-80c296412a6b {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1442.243443] env[63088]: DEBUG oslo_vmware.api [-] Task: {'id': task-1285700, 'name': CreateVM_Task, 'duration_secs': 0.284117} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1442.243798] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] Created VM on the ESX host {{(pid=63088) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1442.250608] env[63088]: DEBUG oslo_concurrency.lockutils [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1442.250785] env[63088]: DEBUG oslo_concurrency.lockutils [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1442.251160] env[63088]: DEBUG oslo_concurrency.lockutils [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1442.251428] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4bdb8826-3fc8-4be3-abf0-b666845ee0d1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1442.255830] env[63088]: DEBUG oslo_vmware.api [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Waiting for the task: (returnval){ [ 1442.255830] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]528785cd-91ee-a90f-e7fe-c21d59a5f0b1" [ 1442.255830] env[63088]: _type = "Task" [ 1442.255830] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1442.263558] env[63088]: DEBUG oslo_vmware.api [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]528785cd-91ee-a90f-e7fe-c21d59a5f0b1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1442.676640] env[63088]: DEBUG nova.network.neutron [req-511e4ae3-abc9-4506-8da8-17ff98f5d0ab req-738ec0ec-084e-4baf-a427-ef6075ab850d service nova] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] Updated VIF entry in instance network info cache for port ab8e9f5f-b27c-4b87-a4a4-80c296412a6b. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1442.677017] env[63088]: DEBUG nova.network.neutron [req-511e4ae3-abc9-4506-8da8-17ff98f5d0ab req-738ec0ec-084e-4baf-a427-ef6075ab850d service nova] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] Updating instance_info_cache with network_info: [{"id": "ab8e9f5f-b27c-4b87-a4a4-80c296412a6b", "address": "fa:16:3e:7f:9c:6b", "network": {"id": "ab40c372-54e5-49d6-a02a-279901071b67", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1152691158-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e4d8e091fab04519a6c430290830998c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapab8e9f5f-b2", "ovs_interfaceid": "ab8e9f5f-b27c-4b87-a4a4-80c296412a6b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1442.765723] env[63088]: DEBUG oslo_vmware.api [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]528785cd-91ee-a90f-e7fe-c21d59a5f0b1, 'name': SearchDatastore_Task, 'duration_secs': 0.009607} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1442.766041] env[63088]: DEBUG oslo_concurrency.lockutils [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1442.766288] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] Processing image 1e8c5d18-0a03-4e18-afe1-de5a6e255953 {{(pid=63088) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1442.766560] env[63088]: DEBUG oslo_concurrency.lockutils [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1442.766718] env[63088]: DEBUG oslo_concurrency.lockutils [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1442.766900] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1442.767177] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1cd4bb53-881e-4d2a-a97c-65a3fff93da9 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1442.775031] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63088) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1442.775216] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63088) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1442.775928] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7fa3ecf2-73f5-4bf4-91ef-440e81c5644d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1442.780587] env[63088]: DEBUG oslo_vmware.api [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Waiting for the task: (returnval){ [ 1442.780587] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52fea6cc-6405-dcc3-f742-ea87d428288b" [ 1442.780587] env[63088]: _type = "Task" [ 1442.780587] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1442.788914] env[63088]: DEBUG oslo_vmware.api [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52fea6cc-6405-dcc3-f742-ea87d428288b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1443.179443] env[63088]: DEBUG oslo_concurrency.lockutils [req-511e4ae3-abc9-4506-8da8-17ff98f5d0ab req-738ec0ec-084e-4baf-a427-ef6075ab850d service nova] Releasing lock "refresh_cache-7af8d406-b000-4e0c-a230-69d114672eb5" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1443.291011] env[63088]: DEBUG oslo_vmware.api [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52fea6cc-6405-dcc3-f742-ea87d428288b, 'name': SearchDatastore_Task, 'duration_secs': 0.008137} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1443.291782] env[63088]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1cddf48c-b6d6-4de2-9847-d3eeff8eb4b3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1443.296489] env[63088]: DEBUG oslo_vmware.api [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Waiting for the task: (returnval){ [ 1443.296489] env[63088]: value = "session[521893a1-57d1-cdb5-ae89-395b16ec493c]52ffb685-621b-defe-628a-506b2d55a007" [ 1443.296489] env[63088]: _type = "Task" [ 1443.296489] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1443.303820] env[63088]: DEBUG oslo_vmware.api [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52ffb685-621b-defe-628a-506b2d55a007, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1443.806237] env[63088]: DEBUG oslo_vmware.api [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': session[521893a1-57d1-cdb5-ae89-395b16ec493c]52ffb685-621b-defe-628a-506b2d55a007, 'name': SearchDatastore_Task, 'duration_secs': 0.010071} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1443.806532] env[63088]: DEBUG oslo_concurrency.lockutils [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1443.806807] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] 7af8d406-b000-4e0c-a230-69d114672eb5/7af8d406-b000-4e0c-a230-69d114672eb5.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1443.807083] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f1b7f2d2-b93b-4450-9b3e-d0dadca3fabb {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1443.813490] env[63088]: DEBUG oslo_vmware.api [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Waiting for the task: (returnval){ [ 1443.813490] env[63088]: value = "task-1285701" [ 1443.813490] env[63088]: _type = "Task" [ 1443.813490] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1443.820756] env[63088]: DEBUG oslo_vmware.api [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285701, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1444.322635] env[63088]: DEBUG oslo_vmware.api [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285701, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.431514} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1444.322994] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1e8c5d18-0a03-4e18-afe1-de5a6e255953/1e8c5d18-0a03-4e18-afe1-de5a6e255953.vmdk to [datastore1] 7af8d406-b000-4e0c-a230-69d114672eb5/7af8d406-b000-4e0c-a230-69d114672eb5.vmdk {{(pid=63088) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1444.323121] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] Extending root virtual disk to 1048576 {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1444.323376] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e2776205-375b-47a4-91b1-1b737a13d8da {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1444.329273] env[63088]: DEBUG oslo_vmware.api [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Waiting for the task: (returnval){ [ 1444.329273] env[63088]: value = "task-1285702" [ 1444.329273] env[63088]: _type = "Task" [ 1444.329273] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1444.336714] env[63088]: DEBUG oslo_vmware.api [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285702, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1444.838857] env[63088]: DEBUG oslo_vmware.api [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285702, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.057271} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1444.839148] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] Extended root virtual disk {{(pid=63088) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1444.839906] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44fde092-c6bb-4e50-ae3d-8f3c349239bb {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1444.860862] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] Reconfiguring VM instance instance-00000070 to attach disk [datastore1] 7af8d406-b000-4e0c-a230-69d114672eb5/7af8d406-b000-4e0c-a230-69d114672eb5.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1444.861103] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f900b9ce-7695-44da-a397-65517f9a7f70 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1444.879720] env[63088]: DEBUG oslo_vmware.api [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Waiting for the task: (returnval){ [ 1444.879720] env[63088]: value = "task-1285703" [ 1444.879720] env[63088]: _type = "Task" [ 1444.879720] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1444.886550] env[63088]: DEBUG oslo_vmware.api [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285703, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1445.390582] env[63088]: DEBUG oslo_vmware.api [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285703, 'name': ReconfigVM_Task, 'duration_secs': 0.261141} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1445.390951] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] Reconfigured VM instance instance-00000070 to attach disk [datastore1] 7af8d406-b000-4e0c-a230-69d114672eb5/7af8d406-b000-4e0c-a230-69d114672eb5.vmdk or device None with type sparse {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1445.391473] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2180e6e8-1280-4c6d-8ac2-05176c74b56d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1445.397831] env[63088]: DEBUG oslo_vmware.api [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Waiting for the task: (returnval){ [ 1445.397831] env[63088]: value = "task-1285704" [ 1445.397831] env[63088]: _type = "Task" [ 1445.397831] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1445.405081] env[63088]: DEBUG oslo_vmware.api [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285704, 'name': Rename_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1445.907475] env[63088]: DEBUG oslo_vmware.api [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285704, 'name': Rename_Task, 'duration_secs': 0.123444} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1445.907788] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] Powering on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1445.907952] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ad4a1e89-5f2d-46f3-ad48-ef555e6f2a15 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1445.913703] env[63088]: DEBUG oslo_vmware.api [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Waiting for the task: (returnval){ [ 1445.913703] env[63088]: value = "task-1285705" [ 1445.913703] env[63088]: _type = "Task" [ 1445.913703] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1445.920639] env[63088]: DEBUG oslo_vmware.api [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285705, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1446.425144] env[63088]: DEBUG oslo_vmware.api [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285705, 'name': PowerOnVM_Task, 'duration_secs': 0.407729} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1446.425538] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] Powered on the VM {{(pid=63088) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1446.425696] env[63088]: INFO nova.compute.manager [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] Took 6.61 seconds to spawn the instance on the hypervisor. [ 1446.425924] env[63088]: DEBUG nova.compute.manager [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] Checking state {{(pid=63088) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1446.426715] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-007fbd33-9b0d-4aa8-8c9b-743282ff776d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1446.943615] env[63088]: INFO nova.compute.manager [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] Took 11.27 seconds to build instance. [ 1447.444786] env[63088]: DEBUG oslo_concurrency.lockutils [None req-ecba7a82-71a1-4c67-ad81-a67470641461 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Lock "7af8d406-b000-4e0c-a230-69d114672eb5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.778s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1447.733156] env[63088]: DEBUG nova.compute.manager [req-9dcb8475-69ee-4234-b420-1a9b64c8ab9f req-3e20eb36-9f40-41d8-88a6-e003139cb7bf service nova] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] Received event network-changed-ab8e9f5f-b27c-4b87-a4a4-80c296412a6b {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1447.733298] env[63088]: DEBUG nova.compute.manager [req-9dcb8475-69ee-4234-b420-1a9b64c8ab9f req-3e20eb36-9f40-41d8-88a6-e003139cb7bf service nova] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] Refreshing instance network info cache due to event network-changed-ab8e9f5f-b27c-4b87-a4a4-80c296412a6b. {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1447.733516] env[63088]: DEBUG oslo_concurrency.lockutils [req-9dcb8475-69ee-4234-b420-1a9b64c8ab9f req-3e20eb36-9f40-41d8-88a6-e003139cb7bf service nova] Acquiring lock "refresh_cache-7af8d406-b000-4e0c-a230-69d114672eb5" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1447.733661] env[63088]: DEBUG oslo_concurrency.lockutils [req-9dcb8475-69ee-4234-b420-1a9b64c8ab9f req-3e20eb36-9f40-41d8-88a6-e003139cb7bf service nova] Acquired lock "refresh_cache-7af8d406-b000-4e0c-a230-69d114672eb5" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1447.733824] env[63088]: DEBUG nova.network.neutron [req-9dcb8475-69ee-4234-b420-1a9b64c8ab9f req-3e20eb36-9f40-41d8-88a6-e003139cb7bf service nova] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] Refreshing network info cache for port ab8e9f5f-b27c-4b87-a4a4-80c296412a6b {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1448.432599] env[63088]: DEBUG nova.network.neutron [req-9dcb8475-69ee-4234-b420-1a9b64c8ab9f req-3e20eb36-9f40-41d8-88a6-e003139cb7bf service nova] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] Updated VIF entry in instance network info cache for port ab8e9f5f-b27c-4b87-a4a4-80c296412a6b. {{(pid=63088) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1448.432968] env[63088]: DEBUG nova.network.neutron [req-9dcb8475-69ee-4234-b420-1a9b64c8ab9f req-3e20eb36-9f40-41d8-88a6-e003139cb7bf service nova] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] Updating instance_info_cache with network_info: [{"id": "ab8e9f5f-b27c-4b87-a4a4-80c296412a6b", "address": "fa:16:3e:7f:9c:6b", "network": {"id": "ab40c372-54e5-49d6-a02a-279901071b67", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1152691158-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.249", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e4d8e091fab04519a6c430290830998c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapab8e9f5f-b2", "ovs_interfaceid": "ab8e9f5f-b27c-4b87-a4a4-80c296412a6b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1448.936249] env[63088]: DEBUG oslo_concurrency.lockutils [req-9dcb8475-69ee-4234-b420-1a9b64c8ab9f req-3e20eb36-9f40-41d8-88a6-e003139cb7bf service nova] Releasing lock "refresh_cache-7af8d406-b000-4e0c-a230-69d114672eb5" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1456.106643] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1459.604950] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1462.604472] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1463.605245] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1463.605614] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63088) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1465.604815] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1465.605228] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Starting heal instance info cache {{(pid=63088) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1465.605228] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Rebuilding the list of instances to heal {{(pid=63088) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 1466.178425] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Acquiring lock "refresh_cache-7af8d406-b000-4e0c-a230-69d114672eb5" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1466.178589] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Acquired lock "refresh_cache-7af8d406-b000-4e0c-a230-69d114672eb5" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1466.178765] env[63088]: DEBUG nova.network.neutron [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] Forcefully refreshing network info cache for instance {{(pid=63088) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1466.178928] env[63088]: DEBUG nova.objects.instance [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Lazy-loading 'info_cache' on Instance uuid 7af8d406-b000-4e0c-a230-69d114672eb5 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1467.895462] env[63088]: DEBUG nova.network.neutron [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] Updating instance_info_cache with network_info: [{"id": "ab8e9f5f-b27c-4b87-a4a4-80c296412a6b", "address": "fa:16:3e:7f:9c:6b", "network": {"id": "ab40c372-54e5-49d6-a02a-279901071b67", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1152691158-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.249", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e4d8e091fab04519a6c430290830998c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapab8e9f5f-b2", "ovs_interfaceid": "ab8e9f5f-b27c-4b87-a4a4-80c296412a6b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1468.398785] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Releasing lock "refresh_cache-7af8d406-b000-4e0c-a230-69d114672eb5" {{(pid=63088) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1468.399055] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] Updated the network info_cache for instance {{(pid=63088) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 1468.399304] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1468.604675] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1468.604957] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1468.605205] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager.update_available_resource {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1469.108587] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1469.108976] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1469.109052] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1469.109180] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63088) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1469.110108] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ed5f642-ecdc-4048-8e21-bbeeaaea3fe0 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1469.118198] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03459af6-eee1-4e10-a0cb-97f4aa6506db {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1469.131759] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88a9b5c5-c9a7-4f93-bd48-ef4eab089974 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1469.137697] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a0dc168-32dc-4e93-8369-ef273adb094f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1469.166954] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181369MB free_disk=140GB free_vcpus=48 pci_devices=None {{(pid=63088) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1469.168672] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1469.168672] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1470.190835] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Instance 7af8d406-b000-4e0c-a230-69d114672eb5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63088) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1470.191232] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=63088) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1470.191232] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=63088) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1470.216259] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42030dcc-553f-4706-aabf-3e6d21e3b9c3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1470.223606] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6fe9378-d46e-4799-9e5a-dc060d5cfe31 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1470.253370] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d55d0a3-1dd6-43b6-8c9b-2f09ea3d0676 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1470.260143] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec44b21d-8b47-4029-bdf1-a3953175530c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1470.272842] env[63088]: DEBUG nova.compute.provider_tree [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1470.775640] env[63088]: DEBUG nova.scheduler.client.report [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1471.280888] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63088) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1471.281321] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.114s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1476.276335] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1486.046591] env[63088]: DEBUG oslo_concurrency.lockutils [None req-7b004600-d256-4b24-8c73-aea11361c1bb tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Acquiring lock "7af8d406-b000-4e0c-a230-69d114672eb5" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1486.046890] env[63088]: DEBUG oslo_concurrency.lockutils [None req-7b004600-d256-4b24-8c73-aea11361c1bb tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Lock "7af8d406-b000-4e0c-a230-69d114672eb5" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1486.550686] env[63088]: DEBUG nova.compute.utils [None req-7b004600-d256-4b24-8c73-aea11361c1bb tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Using /dev/sd instead of None {{(pid=63088) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1487.053609] env[63088]: DEBUG oslo_concurrency.lockutils [None req-7b004600-d256-4b24-8c73-aea11361c1bb tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Lock "7af8d406-b000-4e0c-a230-69d114672eb5" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1488.110283] env[63088]: DEBUG oslo_concurrency.lockutils [None req-7b004600-d256-4b24-8c73-aea11361c1bb tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Acquiring lock "7af8d406-b000-4e0c-a230-69d114672eb5" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1488.110677] env[63088]: DEBUG oslo_concurrency.lockutils [None req-7b004600-d256-4b24-8c73-aea11361c1bb tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Lock "7af8d406-b000-4e0c-a230-69d114672eb5" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1488.110778] env[63088]: INFO nova.compute.manager [None req-7b004600-d256-4b24-8c73-aea11361c1bb tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] Attaching volume 74765d2e-ca57-4778-be67-69cbe613ce3a to /dev/sdb [ 1488.139671] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a25b7c5d-4c53-44e6-baa6-af22a6d65f7d {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1488.146679] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45083d0b-e02b-47ba-80f4-9ed3f0eb55fa {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1488.159178] env[63088]: DEBUG nova.virt.block_device [None req-7b004600-d256-4b24-8c73-aea11361c1bb tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] Updating existing volume attachment record: a80eead6-3883-469c-a689-70c821565058 {{(pid=63088) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1492.700957] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-7b004600-d256-4b24-8c73-aea11361c1bb tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] Volume attach. Driver type: vmdk {{(pid=63088) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1492.701236] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-7b004600-d256-4b24-8c73-aea11361c1bb tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-275989', 'volume_id': '74765d2e-ca57-4778-be67-69cbe613ce3a', 'name': 'volume-74765d2e-ca57-4778-be67-69cbe613ce3a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7af8d406-b000-4e0c-a230-69d114672eb5', 'attached_at': '', 'detached_at': '', 'volume_id': '74765d2e-ca57-4778-be67-69cbe613ce3a', 'serial': '74765d2e-ca57-4778-be67-69cbe613ce3a'} {{(pid=63088) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1492.702122] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09d80b76-5a92-4a9b-9f37-4b22b36d6ff3 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1492.718421] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d371094-f789-40ed-9327-c0e581a47f52 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1492.742014] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-7b004600-d256-4b24-8c73-aea11361c1bb tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] Reconfiguring VM instance instance-00000070 to attach disk [datastore1] volume-74765d2e-ca57-4778-be67-69cbe613ce3a/volume-74765d2e-ca57-4778-be67-69cbe613ce3a.vmdk or device None with type thin {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1492.742254] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ba793f61-a452-4e42-846d-c8d401d598fe {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1492.759834] env[63088]: DEBUG oslo_vmware.api [None req-7b004600-d256-4b24-8c73-aea11361c1bb tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Waiting for the task: (returnval){ [ 1492.759834] env[63088]: value = "task-1285708" [ 1492.759834] env[63088]: _type = "Task" [ 1492.759834] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1492.767357] env[63088]: DEBUG oslo_vmware.api [None req-7b004600-d256-4b24-8c73-aea11361c1bb tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285708, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1493.269121] env[63088]: DEBUG oslo_vmware.api [None req-7b004600-d256-4b24-8c73-aea11361c1bb tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285708, 'name': ReconfigVM_Task, 'duration_secs': 0.318529} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1493.269423] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-7b004600-d256-4b24-8c73-aea11361c1bb tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] Reconfigured VM instance instance-00000070 to attach disk [datastore1] volume-74765d2e-ca57-4778-be67-69cbe613ce3a/volume-74765d2e-ca57-4778-be67-69cbe613ce3a.vmdk or device None with type thin {{(pid=63088) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1493.274032] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-aed47136-c19f-48dd-a333-c252921c4e0f {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1493.288844] env[63088]: DEBUG oslo_vmware.api [None req-7b004600-d256-4b24-8c73-aea11361c1bb tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Waiting for the task: (returnval){ [ 1493.288844] env[63088]: value = "task-1285709" [ 1493.288844] env[63088]: _type = "Task" [ 1493.288844] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1493.296380] env[63088]: DEBUG oslo_vmware.api [None req-7b004600-d256-4b24-8c73-aea11361c1bb tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285709, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1493.798807] env[63088]: DEBUG oslo_vmware.api [None req-7b004600-d256-4b24-8c73-aea11361c1bb tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285709, 'name': ReconfigVM_Task, 'duration_secs': 0.163648} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1493.799115] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-7b004600-d256-4b24-8c73-aea11361c1bb tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-275989', 'volume_id': '74765d2e-ca57-4778-be67-69cbe613ce3a', 'name': 'volume-74765d2e-ca57-4778-be67-69cbe613ce3a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7af8d406-b000-4e0c-a230-69d114672eb5', 'attached_at': '', 'detached_at': '', 'volume_id': '74765d2e-ca57-4778-be67-69cbe613ce3a', 'serial': '74765d2e-ca57-4778-be67-69cbe613ce3a'} {{(pid=63088) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1494.834548] env[63088]: DEBUG nova.objects.instance [None req-7b004600-d256-4b24-8c73-aea11361c1bb tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Lazy-loading 'flavor' on Instance uuid 7af8d406-b000-4e0c-a230-69d114672eb5 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1495.341519] env[63088]: DEBUG oslo_concurrency.lockutils [None req-7b004600-d256-4b24-8c73-aea11361c1bb tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Lock "7af8d406-b000-4e0c-a230-69d114672eb5" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.231s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1495.512323] env[63088]: DEBUG oslo_concurrency.lockutils [None req-26bdcfe9-dd79-4ff6-b1b4-71108f471237 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Acquiring lock "7af8d406-b000-4e0c-a230-69d114672eb5" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1495.512581] env[63088]: DEBUG oslo_concurrency.lockutils [None req-26bdcfe9-dd79-4ff6-b1b4-71108f471237 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Lock "7af8d406-b000-4e0c-a230-69d114672eb5" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1496.015955] env[63088]: INFO nova.compute.manager [None req-26bdcfe9-dd79-4ff6-b1b4-71108f471237 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] Detaching volume 74765d2e-ca57-4778-be67-69cbe613ce3a [ 1496.046161] env[63088]: INFO nova.virt.block_device [None req-26bdcfe9-dd79-4ff6-b1b4-71108f471237 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] Attempting to driver detach volume 74765d2e-ca57-4778-be67-69cbe613ce3a from mountpoint /dev/sdb [ 1496.046427] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-26bdcfe9-dd79-4ff6-b1b4-71108f471237 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] Volume detach. Driver type: vmdk {{(pid=63088) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1496.046619] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-26bdcfe9-dd79-4ff6-b1b4-71108f471237 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-275989', 'volume_id': '74765d2e-ca57-4778-be67-69cbe613ce3a', 'name': 'volume-74765d2e-ca57-4778-be67-69cbe613ce3a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7af8d406-b000-4e0c-a230-69d114672eb5', 'attached_at': '', 'detached_at': '', 'volume_id': '74765d2e-ca57-4778-be67-69cbe613ce3a', 'serial': '74765d2e-ca57-4778-be67-69cbe613ce3a'} {{(pid=63088) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1496.047545] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b7669b6-01d2-472f-821c-ea347624c6f2 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1496.068694] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30c4fb66-3f6c-4077-a346-ce557c897808 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1496.075016] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32b7c5f1-f944-472a-bff5-93853386ed5c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1496.094226] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35ba3418-4f7d-404f-a5d2-95d11cd20970 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1496.107903] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-26bdcfe9-dd79-4ff6-b1b4-71108f471237 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] The volume has not been displaced from its original location: [datastore1] volume-74765d2e-ca57-4778-be67-69cbe613ce3a/volume-74765d2e-ca57-4778-be67-69cbe613ce3a.vmdk. No consolidation needed. {{(pid=63088) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1496.113057] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-26bdcfe9-dd79-4ff6-b1b4-71108f471237 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] Reconfiguring VM instance instance-00000070 to detach disk 2001 {{(pid=63088) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1496.113353] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ef1082c4-3f54-43a0-a87c-a10976e8bfdb {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1496.130487] env[63088]: DEBUG oslo_vmware.api [None req-26bdcfe9-dd79-4ff6-b1b4-71108f471237 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Waiting for the task: (returnval){ [ 1496.130487] env[63088]: value = "task-1285710" [ 1496.130487] env[63088]: _type = "Task" [ 1496.130487] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1496.140836] env[63088]: DEBUG oslo_vmware.api [None req-26bdcfe9-dd79-4ff6-b1b4-71108f471237 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285710, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1496.641757] env[63088]: DEBUG oslo_vmware.api [None req-26bdcfe9-dd79-4ff6-b1b4-71108f471237 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285710, 'name': ReconfigVM_Task, 'duration_secs': 0.195624} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1496.642090] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-26bdcfe9-dd79-4ff6-b1b4-71108f471237 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] Reconfigured VM instance instance-00000070 to detach disk 2001 {{(pid=63088) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1496.646513] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-045fa23b-7cc4-4456-aa9d-1d501e7fe873 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1496.661233] env[63088]: DEBUG oslo_vmware.api [None req-26bdcfe9-dd79-4ff6-b1b4-71108f471237 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Waiting for the task: (returnval){ [ 1496.661233] env[63088]: value = "task-1285711" [ 1496.661233] env[63088]: _type = "Task" [ 1496.661233] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1496.668637] env[63088]: DEBUG oslo_vmware.api [None req-26bdcfe9-dd79-4ff6-b1b4-71108f471237 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285711, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1497.170164] env[63088]: DEBUG oslo_vmware.api [None req-26bdcfe9-dd79-4ff6-b1b4-71108f471237 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285711, 'name': ReconfigVM_Task, 'duration_secs': 0.149024} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1497.170494] env[63088]: DEBUG nova.virt.vmwareapi.volumeops [None req-26bdcfe9-dd79-4ff6-b1b4-71108f471237 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-275989', 'volume_id': '74765d2e-ca57-4778-be67-69cbe613ce3a', 'name': 'volume-74765d2e-ca57-4778-be67-69cbe613ce3a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7af8d406-b000-4e0c-a230-69d114672eb5', 'attached_at': '', 'detached_at': '', 'volume_id': '74765d2e-ca57-4778-be67-69cbe613ce3a', 'serial': '74765d2e-ca57-4778-be67-69cbe613ce3a'} {{(pid=63088) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1497.709805] env[63088]: DEBUG nova.objects.instance [None req-26bdcfe9-dd79-4ff6-b1b4-71108f471237 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Lazy-loading 'flavor' on Instance uuid 7af8d406-b000-4e0c-a230-69d114672eb5 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1498.718315] env[63088]: DEBUG oslo_concurrency.lockutils [None req-26bdcfe9-dd79-4ff6-b1b4-71108f471237 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Lock "7af8d406-b000-4e0c-a230-69d114672eb5" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.205s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1499.740311] env[63088]: DEBUG oslo_concurrency.lockutils [None req-e2a9d4bd-ad89-4cd6-8620-b95a3250fe85 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Acquiring lock "7af8d406-b000-4e0c-a230-69d114672eb5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1499.740719] env[63088]: DEBUG oslo_concurrency.lockutils [None req-e2a9d4bd-ad89-4cd6-8620-b95a3250fe85 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Lock "7af8d406-b000-4e0c-a230-69d114672eb5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1499.740827] env[63088]: DEBUG oslo_concurrency.lockutils [None req-e2a9d4bd-ad89-4cd6-8620-b95a3250fe85 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Acquiring lock "7af8d406-b000-4e0c-a230-69d114672eb5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1499.741018] env[63088]: DEBUG oslo_concurrency.lockutils [None req-e2a9d4bd-ad89-4cd6-8620-b95a3250fe85 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Lock "7af8d406-b000-4e0c-a230-69d114672eb5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1499.741208] env[63088]: DEBUG oslo_concurrency.lockutils [None req-e2a9d4bd-ad89-4cd6-8620-b95a3250fe85 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Lock "7af8d406-b000-4e0c-a230-69d114672eb5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1499.743529] env[63088]: INFO nova.compute.manager [None req-e2a9d4bd-ad89-4cd6-8620-b95a3250fe85 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] Terminating instance [ 1499.745364] env[63088]: DEBUG nova.compute.manager [None req-e2a9d4bd-ad89-4cd6-8620-b95a3250fe85 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] Start destroying the instance on the hypervisor. {{(pid=63088) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1499.745564] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-e2a9d4bd-ad89-4cd6-8620-b95a3250fe85 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] Destroying instance {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1499.746482] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7f3e4a0-c612-4546-921d-383b8d511932 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1499.754446] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2a9d4bd-ad89-4cd6-8620-b95a3250fe85 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] Powering off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1499.754670] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0e3cc2a9-7835-4d4b-87d1-7201a1d77c07 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1499.760828] env[63088]: DEBUG oslo_vmware.api [None req-e2a9d4bd-ad89-4cd6-8620-b95a3250fe85 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Waiting for the task: (returnval){ [ 1499.760828] env[63088]: value = "task-1285712" [ 1499.760828] env[63088]: _type = "Task" [ 1499.760828] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1499.768843] env[63088]: DEBUG oslo_vmware.api [None req-e2a9d4bd-ad89-4cd6-8620-b95a3250fe85 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285712, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1500.271320] env[63088]: DEBUG oslo_vmware.api [None req-e2a9d4bd-ad89-4cd6-8620-b95a3250fe85 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285712, 'name': PowerOffVM_Task, 'duration_secs': 0.18317} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1500.271587] env[63088]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2a9d4bd-ad89-4cd6-8620-b95a3250fe85 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] Powered off the VM {{(pid=63088) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1500.271765] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-e2a9d4bd-ad89-4cd6-8620-b95a3250fe85 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] Unregistering the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1500.272024] env[63088]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ce3e1823-3b5c-4531-9092-fc602723d330 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1500.332195] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-e2a9d4bd-ad89-4cd6-8620-b95a3250fe85 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] Unregistered the VM {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1500.332442] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-e2a9d4bd-ad89-4cd6-8620-b95a3250fe85 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] Deleting contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1500.332615] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-e2a9d4bd-ad89-4cd6-8620-b95a3250fe85 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Deleting the datastore file [datastore1] 7af8d406-b000-4e0c-a230-69d114672eb5 {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1500.332876] env[63088]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a6a3ac8e-d59a-4e57-a029-00367b976fe1 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1500.338450] env[63088]: DEBUG oslo_vmware.api [None req-e2a9d4bd-ad89-4cd6-8620-b95a3250fe85 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Waiting for the task: (returnval){ [ 1500.338450] env[63088]: value = "task-1285714" [ 1500.338450] env[63088]: _type = "Task" [ 1500.338450] env[63088]: } to complete. {{(pid=63088) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1500.345815] env[63088]: DEBUG oslo_vmware.api [None req-e2a9d4bd-ad89-4cd6-8620-b95a3250fe85 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285714, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1500.848614] env[63088]: DEBUG oslo_vmware.api [None req-e2a9d4bd-ad89-4cd6-8620-b95a3250fe85 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Task: {'id': task-1285714, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.13682} completed successfully. {{(pid=63088) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1500.848952] env[63088]: DEBUG nova.virt.vmwareapi.ds_util [None req-e2a9d4bd-ad89-4cd6-8620-b95a3250fe85 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Deleted the datastore file {{(pid=63088) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1500.849082] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-e2a9d4bd-ad89-4cd6-8620-b95a3250fe85 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] Deleted contents of the VM from datastore datastore1 {{(pid=63088) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1500.849275] env[63088]: DEBUG nova.virt.vmwareapi.vmops [None req-e2a9d4bd-ad89-4cd6-8620-b95a3250fe85 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] Instance destroyed {{(pid=63088) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1500.849448] env[63088]: INFO nova.compute.manager [None req-e2a9d4bd-ad89-4cd6-8620-b95a3250fe85 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1500.849691] env[63088]: DEBUG oslo.service.loopingcall [None req-e2a9d4bd-ad89-4cd6-8620-b95a3250fe85 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63088) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1500.849885] env[63088]: DEBUG nova.compute.manager [-] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] Deallocating network for instance {{(pid=63088) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1500.849980] env[63088]: DEBUG nova.network.neutron [-] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] deallocate_for_instance() {{(pid=63088) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1501.339356] env[63088]: DEBUG nova.compute.manager [req-1266948e-70bd-411e-9e70-4631415e0570 req-28168e32-fb2f-4a3f-b4e6-a69c2b0fb1ba service nova] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] Received event network-vif-deleted-ab8e9f5f-b27c-4b87-a4a4-80c296412a6b {{(pid=63088) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1501.339570] env[63088]: INFO nova.compute.manager [req-1266948e-70bd-411e-9e70-4631415e0570 req-28168e32-fb2f-4a3f-b4e6-a69c2b0fb1ba service nova] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] Neutron deleted interface ab8e9f5f-b27c-4b87-a4a4-80c296412a6b; detaching it from the instance and deleting it from the info cache [ 1501.339751] env[63088]: DEBUG nova.network.neutron [req-1266948e-70bd-411e-9e70-4631415e0570 req-28168e32-fb2f-4a3f-b4e6-a69c2b0fb1ba service nova] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1501.747355] env[63088]: DEBUG nova.network.neutron [-] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] Updating instance_info_cache with network_info: [] {{(pid=63088) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1501.842881] env[63088]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-de0e273b-34cd-4e84-804c-68cd382aa753 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1501.853315] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba1d3fbf-013f-449f-8c6b-d0af62101d61 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1501.875213] env[63088]: DEBUG nova.compute.manager [req-1266948e-70bd-411e-9e70-4631415e0570 req-28168e32-fb2f-4a3f-b4e6-a69c2b0fb1ba service nova] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] Detach interface failed, port_id=ab8e9f5f-b27c-4b87-a4a4-80c296412a6b, reason: Instance 7af8d406-b000-4e0c-a230-69d114672eb5 could not be found. {{(pid=63088) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1502.250415] env[63088]: INFO nova.compute.manager [-] [instance: 7af8d406-b000-4e0c-a230-69d114672eb5] Took 1.40 seconds to deallocate network for instance. [ 1502.757438] env[63088]: DEBUG oslo_concurrency.lockutils [None req-e2a9d4bd-ad89-4cd6-8620-b95a3250fe85 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1502.757719] env[63088]: DEBUG oslo_concurrency.lockutils [None req-e2a9d4bd-ad89-4cd6-8620-b95a3250fe85 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1502.757963] env[63088]: DEBUG nova.objects.instance [None req-e2a9d4bd-ad89-4cd6-8620-b95a3250fe85 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Lazy-loading 'resources' on Instance uuid 7af8d406-b000-4e0c-a230-69d114672eb5 {{(pid=63088) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1503.279943] env[63088]: DEBUG nova.scheduler.client.report [None req-e2a9d4bd-ad89-4cd6-8620-b95a3250fe85 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Refreshing inventories for resource provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1503.293544] env[63088]: DEBUG nova.scheduler.client.report [None req-e2a9d4bd-ad89-4cd6-8620-b95a3250fe85 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Updating ProviderTree inventory for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1503.293788] env[63088]: DEBUG nova.compute.provider_tree [None req-e2a9d4bd-ad89-4cd6-8620-b95a3250fe85 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Updating inventory in ProviderTree for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1503.303955] env[63088]: DEBUG nova.scheduler.client.report [None req-e2a9d4bd-ad89-4cd6-8620-b95a3250fe85 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Refreshing aggregate associations for resource provider 6eae54a9-8831-40eb-bf54-4bc60d346b02, aggregates: None {{(pid=63088) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1503.320283] env[63088]: DEBUG nova.scheduler.client.report [None req-e2a9d4bd-ad89-4cd6-8620-b95a3250fe85 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Refreshing trait associations for resource provider 6eae54a9-8831-40eb-bf54-4bc60d346b02, traits: COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO {{(pid=63088) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1503.343709] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e12c6bbe-ba8c-41b2-a9fe-504a38459822 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1503.351046] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a34a092d-59d4-4adc-877f-83e7cd374f7c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1503.379428] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b0877f4-4df7-46c1-9f40-e90b7286e507 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1503.386240] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfc18d3a-f061-4556-a9a0-b0cdb9aed39c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1503.400259] env[63088]: DEBUG nova.compute.provider_tree [None req-e2a9d4bd-ad89-4cd6-8620-b95a3250fe85 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1503.903633] env[63088]: DEBUG nova.scheduler.client.report [None req-e2a9d4bd-ad89-4cd6-8620-b95a3250fe85 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1504.409030] env[63088]: DEBUG oslo_concurrency.lockutils [None req-e2a9d4bd-ad89-4cd6-8620-b95a3250fe85 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.651s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1504.431716] env[63088]: INFO nova.scheduler.client.report [None req-e2a9d4bd-ad89-4cd6-8620-b95a3250fe85 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Deleted allocations for instance 7af8d406-b000-4e0c-a230-69d114672eb5 [ 1504.941040] env[63088]: DEBUG oslo_concurrency.lockutils [None req-e2a9d4bd-ad89-4cd6-8620-b95a3250fe85 tempest-AttachVolumeNegativeTest-229114359 tempest-AttachVolumeNegativeTest-229114359-project-member] Lock "7af8d406-b000-4e0c-a230-69d114672eb5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.200s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1517.604617] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1519.604173] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1523.604491] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1523.605100] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1523.605100] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63088) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1526.604588] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1526.604900] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Starting heal instance info cache {{(pid=63088) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1526.604900] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Rebuilding the list of instances to heal {{(pid=63088) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 1527.107669] env[63088]: DEBUG nova.compute.manager [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Didn't find any instances for network info cache update. {{(pid=63088) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10012}} [ 1527.603990] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1528.600320] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1529.604239] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1529.604659] env[63088]: DEBUG oslo_service.periodic_task [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Running periodic task ComputeManager.update_available_resource {{(pid=63088) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1530.107914] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1530.108220] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1530.108410] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1530.108567] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63088) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1530.109513] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3acdf1e-419c-4142-b361-fc955a158907 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1530.117945] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-127a16c0-7386-481c-9dca-da46a6b2c46c {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1530.131745] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cac6a0c7-daae-42be-a52b-cad6bf35d062 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1530.137848] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d56a89c8-077e-4d6a-ba5c-53842e79535b {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1530.167224] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181477MB free_disk=140GB free_vcpus=48 pci_devices=None {{(pid=63088) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1530.167425] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1530.167606] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1531.186545] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=63088) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1531.186792] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=63088) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1531.199400] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13059e08-0419-4c9a-a066-05fd09396aae {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1531.207083] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e64d224-aaf8-466f-ac16-e1a9880f3176 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1531.235687] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4db4257-5a42-45b2-8f49-dd2ad2675d76 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1531.242286] env[63088]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce6c6de3-0d66-430c-9810-3c57dd0c5e23 {{(pid=63088) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1531.255896] env[63088]: DEBUG nova.compute.provider_tree [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Inventory has not changed in ProviderTree for provider: 6eae54a9-8831-40eb-bf54-4bc60d346b02 {{(pid=63088) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1531.759471] env[63088]: DEBUG nova.scheduler.client.report [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Inventory has not changed for provider 6eae54a9-8831-40eb-bf54-4bc60d346b02 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 140, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63088) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1532.264596] env[63088]: DEBUG nova.compute.resource_tracker [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63088) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1532.264997] env[63088]: DEBUG oslo_concurrency.lockutils [None req-f7d4a5f8-0d16-4b4b-9747-424fed7ba166 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.097s {{(pid=63088) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}}